var/home/core/zuul-output/0000755000175000017500000000000015140137302014521 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015140143771015474 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000234255115140143712020260 0ustar corecoreǀikubelet.lognc9r~DYA6ZF,-K$l"mklkcQӖHSd.~"mv?_eGbuuțx{w7ݭ7֫'e% oo/q3m^]/o?8.7oW}ʋghewx/mX,ojŻ ^Tb3b#׳:}=p7뼝ca㑔`e0I1Q!&ѱ[/o^{W-{t3_U|6 x)K#/5ΌR"ggóisR)N %emOQ/Ϋ[oa0vs68/Jʢ ܚʂ9ss3+aô٥J}{37FEbп3 FKX1QRQlrTvb)E,s)Wɀ;$#LcdHM%vz_. o~I|g\W#NqɌDSd1d9nT#Abn q1J# !8,$RNI? j!bE"o j/o\E`r"hA ós yi\[.!=A(%Ud,QwC}F][UVYE NQGn0Ƞɻ>.ww}(o./WY<͉#5O H 'wo6C9yg|O~ €'} S[q?,!yq%a:y<\tunL h%$Ǥ].v y[W_` \r/Ɛ%aޗ' B.-^ mQYd'xP2ewEڊL|^ͣrZg7n͐AG%ʷr<>; 2W>h?y|(G>ClsXT(VIx$(J:&~CQpkۗgVKx*lJ3o|s`<՛=JPBUGߩnX#;4ٻO2{Fݫr~AreFj?wQC9yO|$UvވkZoIfzC|]|[>ӸUKҳt17ä$ ֈm maUNvS_$qrMY QOΨN!㞊;4U^Z/ QB?q3En.اeI"X#gZ+Xk?povR]8~깮$b@n3xh!|t{: CºC{ 8Ѿm[ ~z/9آs;DPsif39HoN λC?; H^-¸oZ( +"@@%'0MtWG uIo1]ߔr TGGJ\ C.iTEZ{(¥:-³xlՐ0A_Fݗw)(c>/1:N3cl.:f 3 JJ5Z|&הԟ,Tصp&NI%`t3Vi=Ob㸵2*3d*mQ%"h+ "f "D(~~moH|E3*46$Ag4aX)Ǜƾ9U Ӆ^};ڲ7J9@ kV%g>a~W;D=;y|AAY'"葋_d$Ə{(he NSfX1982TH#D֪v3l"<, { Tms'oI&'Adp]{1DL^5"Ϧޙ`F}W5XDV7V5EE9esYYfiMOV i/ f>3VQ 7,oTW⇊AqO:rƭĘ DuZ^ To3dEN/} fI+?|Uz5SUZa{P,97óI,Q{eNFV+(hʺb ״ʻʞX6ýcsT z`q 0C?41- _n^ylSO2|'W'BOTLl-9Ja [$3BV2DC4l!TO C*Mrii1f5 JA *#jv߿Imy%u LOL8c3ilLJ!Ip,2(( *%KGj   %*e5-wFp"a~fzqu6tY,d,`!qIv꜒"T[1!I!NwL}\|}.b3oXR\(L _nJB/_xY.# ſԸv}9U}'/o uSH<:˷tGLS0l/LKcQ.os2% t)Eh~2p cL1%'4-1a_`[Zz㧦|k˭c ĚOρ_} Ewt3th?tvͪ{~;J0= |JUԍ;Iw}/9nh7l%>'ct Հ}a>-:(QxPyA Z UcÖgڌ:8cΗ|U1,-N9 dI [@3YNEgl1$9  ֲQ$'dJVE%mT{z`R$77.N|b>harNJ(Bň0ae3V#b,PY0TEu1L/]MTB4$`H6NI\nbǛ*AyA\(u|@ [h-,j7gDTÎ4oWJ$j!fc̖F4BJ2ᮚ苮p(r%Q 6<$(Ӣ(RvA A-^dX? I,($F{ձ7*Oy 6EK( EF #31J8mN .TTF9㕴/5~RxCe,&v3,JE- ZF5%Da,GܠYr猸p$nu?ݣ RF]NHw2k혿q}lrCy u)xF$Z83Ec罋}[εUX%}< ݻln"sv&{b%^AAoۺ(I#hKD:Bߩ#蘈f=9oN*.Ѓ M#JC1?tean`3-SHq$2[ĜSjXRx?}-m6Mw'yR3q㕐)HW'X1BEb $xd(21i)//_і/Cޮm0VKz>I; >d[5Z=4>5!!T@[4 1.x XF`,?Hh]b-#3J( &uz u8.00-(9ŽZcX Jٯ^蒋*k.\MA/Xp9VqNo}#ƓOފgv[r*hy| IϭR-$$m!-W'wTi:4F5^z3/[{1LK[2nM|[<\t=3^qOp4y}|B}yu}뚬"P.ԘBn방u<#< A Q(j%e1!gkqiP(-ʢ-b7$66|*f\#ߍp{8sx[o%}WS`ýͽ>^U_S1VA20:d T2$47mSl*#lzFP_3yb.63>NKnJۦ^4*rB쑓:5Ǧ٨C.1`mU]+y_:,eXX맻c5ޖSwe݊O4L)69 War)|VϟT;Cq%KK-*i ѩQٰ`DݎGu( 꿢\cXn }7Ҫa nG{Y bcWa?\34 P U!7 _* kTuwmUr%ԀjƮĀdU#^ۈӕ3ΊeBO`^}ܖj49lnAvoI "%\;OF& wctغBܮl##mϸ.6p5k0C5PdKB g:=G<$w 24 6e/!~߽f)Q UbshY5mseڠ5_m@`<ۃ7|qk" L+Y*Ha)j~pu7ި!:E#s:ic.XC^wT/]n2'>^&pnapckL>2QQWo/ݻ<̍8)r`F!Woc0Xq0 R' eQ&Aѣzvw=e&".awfShWjÅD0JkBh]s9Ą)kre_f |Nm8p5H!jR@Aiߒ߈ۥLFTk"5l9O'ϓl5x|_}]#LLb FDK|۟OS栉 A*[Jd >|OvJll9pL RN ]F>}8H0]+ES,n?UU{ x~ʓOy_>/_|&q̑0dd4>vk 60D _o~[Sw3ckpkpLNa ^j 5*<&}kˢmqvۗj=<Tr=[ a^؃ È(<^=xZb [_tܡ&yЋ{ Sym^?̑sU~' Ԓ f\itu)b>5h.=-[ӓI(\?{WƱPz;| \;_D[T/BI GH8@"t*"9QXodՔz q[*ڔC"1Ȋ-R0ڱ}oF4 3vFf#8^Vє+k@ :)@%9@nA B q 62!/ 6G (" u:)fSGAV(e֖t܁ ft~c.!R0N<R{mtdFdHÃФsxBl] " Δ<=9i/ d ␙F9Ґ)Hnxps2wApP!se]I)^ k?'k:%Ѹ)?wɧ6a{r7%]_Ϧi~ԞnZhubW*IakVC-(>Z#"U4Xk1G;7#m eji'ĒGIqB//(O &1I;svHd=mJW~ړUCOīpAiB^MP=MQ`=JB!"]b6Ƞi]ItЀ'Vf:yo=K˞r:( n72-˒#K9T\aVܩO "^OF1%e"xm뻱~0GBeFO0ޑ]w(zM6j\v00ׅYɓHڦd%NzT@gID!EL2$%Ӧ{(gL pWkn\SDKIIKWi^9)N?[tLjV}}O͌:&c!JC{J` nKlȉW$)YLE%I:/8)*H|]}\E$V*#(G;3U-;q7KǰfξC?ke`~UK mtIC8^P߼fub8P銗KDi'U6K×5 .]H<$ ^D'!" b1D8,?tT q lKxDȜOY2S3ҁ%mo(YT\3}sѦoY=-- /IDd6Gs =[F۴'c,QAIٰ9JXOz);B= @%AIt0v[Ƿ&FJE͙A~IQ%iShnMІt.޿>q=$ts,cJZڗOx2c6 .1zҪR "^Q[ TF )㢥M-GicQ\BL(hO7zNa>>'(Kgc{>/MoD8q̒vv73'9pM&jV3=ɹvYƛ{3iψI4Kp5 d2oOgd||K>R1Qzi#f>夑3KմԔ萴%|xyr>ķx>{E>Z4Ӥ͋#+hI{hNZt 9`b˝`yB,Ȍ=6Z" 8L O)&On?7\7ix@ D_P"~GijbɠM&HtpR:4Si גt&ngb9%islԃ)Hc`ebw|Ī Zg_0FRYeO:F)O>UD;;MY,2ڨi"R"*R2s@AK/u5,b#u>cY^*xkJ7C~pۊ ~;ɰ@ՙ.rT?m0:;}d8ۈ ݨW>.[Vhi̒;̥_9$W!p.zu~9x۾vC;kN?WƟ+fx3SuKQqxST Ζ2%?T74a{N8;lr`$pZds=3jwlL Eڲ t|*n8[#yN SrA GYb8ZIaʼn8 #fg3i`F#5N 3q_M]j 8E!@1vցP7!|+R@;HspSI]ڻCZUcg5pDcIϹ,oN-_XI,3\j ]ٟ5~' SuipA!C厐$&k7dmhz/#"݃,YqCL$ڲ`"MUbeT>Xuv~4Le͢ }UVM)[A`b}mcE]LCEg=2ȴcmZ?E*-8nhױ1xR2ϫCya` A y!?h!9yL%VLU2gr26A!4vbSG ]ꧧWp/ &ee *w$-`J\ ptǣC^p#_`{ К8EW>*(D{ٛ,[fnY𱹞M=6&$<,"lX-Ǐ_whaE 98 (oѢ/Р΅ 7ցl6618ł_1/=fu).s¯?.S[{'g=Ҥ):d8h\y6]t1T7IUV:;.1& ,5΀j:<< +Y?58In'bXIǣO{&V\DŽ0,9f O_"[l:h¢8wݓ19\:f6:+ .3}=uvKc ٹeS<>ij(o'ciS<{1$E[nP b?8E'xv[K+E{,Qƙ1*dcs_Z'407|qBOgYU|U--sG8`u! qGYܷw;ȌCPc_|(RaIBKb+{P.T! =ĦiTob d<>SHr][KqWs7ѝBYǭ~RR"p9dFg|K- obY_vM 4>/]e/dy,8!xŋ5 R<^mYo 3c9(F?hXf~TTX)QӅtӚe~=WtX-sJb?U'3X7J4l+Cj%LPFxŰAVG Y%.9Vnd8? ǫjU3k%E)OD:"Ϳ%E)=}l/'O"Q_4ILAٍKK7'lWQVm0c:%UEhZ].1lcazn2ͦ_DQP/2 re%_bR~r9_7*vrv |S.Z!rV%¢EN$i^B^rX؆ z1ǡXtiK`uk&LO./!Z&p:ˏ!_B{{s1>"=b'K=}|+: :8au"N@#=Ugzy]sTv||Aec Xi.gL'—Ʃb4AUqػ< &}BIrwZ\"t%>6ES5oaPqobb,v 2w s1,jX4W->L!NUy*Gݓ KmmlTbc[O`uxOp  |T!|ik3cL_ AvG i\fs$<;uI\XAV{ˍlJsŅjЙNhwfG8>Vڇg18 O3E*dt:|X`Z)|z&V*"9U_R=Wd<)tc(߯)Y]g5>.1C( .K3g&_P9&`|8|Ldl?6o AMҪ1EzyNAtRuxyn\]q_ߍ&zk.)Eu{_rjuWݚ;*6mMq!R{QWR=oVbmyanUn.Uqsy.?W8 r[zW*8nؿ[;vmcoW]"U;gm>?Z֒Z6`!2XY]-Zcp˿˘ɲ}MV<в~!?YXV+lx)RRfb-I7p)3XɯEr^,bfbKJ'@hX><[@ ,&,]$*բk-Yv5 '1T9!(*t 0'b@񲱥-kc6VnR0h& 0Z|ђ8 CGV[4xIIWN?Yt>lf@ Vi`D~ڇŁQLLkY <ZPKoma_u` !>Z;3F\dEB n+0Z ?&s{ 6(E|<ޭLk1Yn(F!%sx]>CTl9"و5 |ݹր|/#.w0ޒx"khD?O`-9C| &8֨O8VH5uH)28 Ǿ-R9~ +#e;U6]aD6Xzqd5y n';)VKL]O@b OIAG Lmc 2;\d˽$Mu>WmCEQuabAJ;`uy-u.M>9VsWٔo RS`S#m8k;(WAXq 8@+S@+' 8U˜z+ZU;=eTtX->9U-q .AV/|\ǔ%&$]1YINJ2]:a0OWvI.O6xMY0/M$ *s5x{gsəL3{$)ՆbG(}1wt!wVf;I&Xi43غgR 6 ݩJ$)}Ta@ nS*X#r#v6*;WJ-_@q.+?DK១btMp1 1Gȩ f,M`,Lr6E} m"8_SK$_#O;V 7=xLOu-ȹ2NKLjp*: 'SasyrFrcC0 ѱ LKV:U} -:U8t[=EAV$=i[mhm"roe5jqf$i>;V0eOޞ4ccc2J1TN.7q;"sդSP) 0v3-)-ٕAg"pZ: "ka+n!e߮lɹL V3Os\ဝ+A= 2䣔AzG\ ` \vc"Kj61O Px"3Pc /' PW*3GX liWv-6W&)cX |]O;C%8@*Z1%8Gk@5^NtY"Fbi8D'+_1&1 7U^k6v읨gQ`LRx+I&s5Www` q:cdʰ H`X;"}B=-/M~C>''1R[sdJm RD3Q{)bJatdq>*Ct/GǍ-`2:u)"\**dPdvc& HwMlF@a5`+F>ΰ-q>0*s%Q)L>$ćYV\dsEGز/:ٕycZtO 2ze31cDB/eWy!A/V4cbpWaPBIpqS<(lȣ'3K?e Z?ڠ8VSZM}pnqL f2D?mzq*a[~;DY〩b𻾋-]f8dBմVs6傊zF"daeY(R+q%sor|.v\sfa:TX%;3Xl= \k>kqBbB;t@/Cԍ)Ga[ r=nl-w/38ѮI*/=2!j\FW+[3=`BZWX Zd>t*Uǖ\*Fu6Y3[yBPj|LcwaIuR;uݷ㺾|47ߍeys=.EinE% 1zY\+͕߬VͭW_겼cazyU1wOw)Ǽn@6 |lk'Z|VZpsqL5 څB}>u)^v~,󿴝} 3+m𢛲Pz_Sp2auQAP*tLnIXA6L7 8UgKdT)*7>p{Pgi-b)>U6IXabPde Ӽ8Ģ8GɄnb'G ֤Mcv4?>HC78NE@UMc8>`TvZ:}O K޵6#ۿ%&*W]̺6NLJ@xM =~IJېLXN}N7F6&ͰFIXUO6ȾL.s`_}Vj$j1ߙk nb\-<}5Rie;,ʴwOql}5Jf1\,2u -aԢ ͝,2GlȾL͢M|-ޚA:szGuFH<~ }mq7$H}5f*›ZE&D'&;ͱB\a^QG4"'i)?'qBOLcX㢚żO%ʞ汼_}\)s}@::/P2l;%e#AwKFܟDr- q,|r[7Ba1eIxtzYREfb[<ٱ6NŶoQw DD7fܳoK}a\n=(FzL'9s* lGy~ :+Y*\e짶CA5bd,2}w'&~Omd2mE&d{q&/" ;6Odw䳒؋}I.R ǡi$;}HNiy^G%CrZfLvn]\+ dp6/LCx{QE~H~wN碘,\4FGިKMF$C!CJ3U|2ۦ]^/os_/<㨬ZQX ~ 85`F7m''PBg{ҟ=7OGF;xgRW*eBW$-4F$P|DIwaw^N&ݫ%\%&Z";IJf$֮ QvW*zwB0EZz|FQP,f+q ckE2 ݈'2vMq\Ϗ-bY8A¸Dg^ ?f*\5Ca_O&by3Cf*3lyK2JQ9B.R1gr'PYQ{}u?U~GRfu]T'l7VQ5EuYMڧϷdU.2G(QhbXjA`񳵃XfΌ)$$:LȬUa~)4P PnĭwiF*Gbŀ_P5;Ob*V[GQ(C;28AGHy`H#xKRK=L;Lt.]06MX8thSx(㘟GM34o#胊AkܚY_\NEIT{k*GTR*X$lK]oDt ɿU JNR8E!i39 8Xb&KyfyyU:yS%lbZKCSmMm=TJ5e$It\0rгi0ZϡᣳtZ?<3uGw>,2|tAan9z@=VB瘴>S񹌚2oVH1og{IX-f%N&ӣ2E QK .ަb vd$>a,ʺlI#RaKrmgl?q`!E!¾TdZ̋f.e²L5&Uu9Pv:nkӞC{%nr) DEo ge? L")#fw@p"{,@=$ȁ9*O!LGG푽<~+jiFT3`LtrkT}KҾM+f)zv>?GFCS?^p~F)YYۍ/儵̒Y{丫vҝ!'L?}Ntܢ ,5.rc3i"$Շ$_ȟ$i>N'gwvlK ?(Jq\6ro ☄Y_(/Oq:&a;?O<>&Ib.q9[@#!3X?T.ipᔜX޾>XEA`0*?e =|wLBU M.it'ХciV LR5y.j(e8=h-ϰ$;{2Dϣzd"4=J}-ԕ&D" iTg4u nj*`D@C  bQd}Nܘwh#‹+F2^2I1IRƕ gJ]$"*X&{.mЂYgso+q&¯h/?!Ѫ#b(s4}>QNTmq{`R*W\lR"Th(/xgĉ8Uq&Œ;)ǀ]+sysxC٨j+ duښ+.9&JaaZ`ڛQeL0:e586B"t7!LTI#u!FMif æ艕 MоFPxGLDՇD|,ՠ ,S`YZ!bxS{ZܿDQr3ΩR; nJiԥȫ{ ,7ؐ=c!s`pkӂmjQ{tR29dښN LӀu,i : VW0Pv5~3$ji.%I)42ǘ& vj*y)e]6Twe3Mk&@X*jp),xJ|\<óųH!8Rxv f%NZ /zW[eTJ6F:!oM`ly&ȘәQ`TAh8SWwpiAKvBϷ)FkjPCc.֤~^ĵtpOP+R'׵vaZ,0ZMBlkŲzZeQ2yI*j0Ӂl@-mRee@: D$ V*49vgjrh*xY&YSc xx$I O+p;(ȵ9ypuͺwa .2.̯+w_|g6$Zt.h.ԘEd>>oi4Um4eFbn#w ׊c!::GRDG{`co p*C8 4϶nm^h,DUy՞q莆<]kC.MF5Vb A峱ecُ%_xQLvzksǺǫ3.ECݙZl]F$K xnSĢ%5=ܨ"uWq0Jc< p}x7m- o4\LTT wlV3a{k\]_ڶI"%1wCqܶˆ%No9x7 ' 3O8-db &bF01}c8YI^0W+['~K`nIF2 Xm0bwܔbc _T5T xtPyvi cU"/Q)`dd!:)Z~RAw4teK\2 K\-2Z4L"\u=.˦Eg wiSC! .3u1ëK݆CPnw* XYU}5cW,QnFݐHCt/TSyJFb{Z2K_ M cr磻߮-u]1à ?fA'(DU])<m=BRx~r4;7&v9c>>2yܙIe}uGwr$dD]z5>.{3 |Hڦ9%vU/Hl]y{gAMYfJ b1j62ݞe.߀0"UB]ኮq)oѪnVfeVl\*Rxhs Lcj⮕ҧ\G s3s̀Yw͍eoG&MwvֆNoo(>o b.|BIA{1 ~|c>۷|B:'DE9|B #܉ғ[`v8 ,G ~3J{ z3F4TC5K^ٝ*ߧ$C$4af>ߤVJϷ(~͘JsE(eԮۑFNXw1Bp"+ٌLⱝ;% ;!?}*,78j){w]8!3tAƽθTApDεRfmY&#³4le2_.-9UmI}zip0 S }fQh>@Tdnz؈ ;7GǙFă0=C{鬭$,i3rϱO"KcRg~fp \EH Ԩ[;[- ]}Mt- $2C1f"ԧ8lo#N2)݄e+!6fn]6FUzz8 % irwG^>^M( HG1գ>UB&=0KkT"NalKu~a?f\ꡏ{ /m@L)BkճWeZpRj~DE {:J-Y=âdҕ'㨬q=.XUkk/+uGD2Ψ  ֻ|T ">;9_|8T}8>y"sc5Vz`9@2cTN8$pr8Yw8ힱf001%!!+;NŅDPy @| nPRvPb\alw`S p01'FFk4!n^txv"(KȀ m xwmލ JBYnusR/Z7n7N2!@ Q"GS(=%G_v"/+:TQNx @nQU3B ^'4k} Pa] i,Z2`$O,pV P?*#;db?CnM n5 .ӉU[q aBm?H@pKn{HaBN9BP#/IhBw)@A"SZ(xS>Yb8*!:], ,vH9GpwWdk0)輌#z4tZRHDA;O deXJ;}{x/Vg\Z7@; iU8M Ʒj jVSfN~KUYk-(cS6qv,xӯhCnf=В-BB'rR$q6MwDcj9B},yD /…U(]ƴsD7qາr02́L=994CYzCPy>IL6-f-kE&QԋЉܞ"?nʄnpDc&ڳy϶eaGoh!B $C1hDtD+`Wo' zFhX_'M&~2u|2dyڿFՒ_ao/>F).]ÅuaS0ŀQZ0F( ?jOdC?͝uLOAgEk`*?(/;z1N)"}\fl"'W$/N_ `8br~%^KlVSX2ߖx@k q\)]n\77'|zqt:owi/f30}w7=nH$70'CIϑ ʦ6J=xpy]VoPMS׳C:HF^R4a@.zIa I1$Xx6N<Θ|=.Ҹ~ڣ^ԋSР4SP\~V꯺7<f;谧S$+6*)*+h.4+-+s\wxwZ%kom3:OT2iz=ghg?Fn-UU~Gƌ>0nZU`xW)Qgrͤm` T5e2.rSqq9X#rߦV!4buI7,zWI(&4L(8c( 2(M|FFQ Sn/!0,r]|gZ{tZxj Y}W۠YC}aWwmmf[vyȼx{~KWpQ8 cxmf=X&(>7 ̡U~W*gO{!lz^@gO>,ӤVVf+5EW|dp<G~NqH{P˽m ƙ_={5Žzi жq 2F$aXH -o]_l\-g^_}O5kc<ʦi񡬟)7(ʔM_R%66_lA`P6WN"э W Zյ&QG<S縸ǩ8(̊<CeC9p?E?տ'yU lG_gxkπڸ HejYʢz+(o7Q^)CC Yg~[ x:|!-.f/ݦ=>A[271Sę 6c~$+z ˃Qo,Tӂ?e6_BE$ 5fh#W j~WDq ׳d_Yg"Q6Q/j Ln-Rc ӂ\ }Áw+|~ez<2z)M2 BY7'g̤J haQW=!n {kΪY(Oγ9d`ٸhrcKdoㆰz{M_ʞUՏ43:L}jIױS"ORLfЩu]^z,mWʞB0ȓ6n*қ r^tda:E;Jer.r\8(4C@zb6)%WlTϝmD(ECf":<&(\Ԫsp4fH$4>8 #Wb */eTA5:%l>#@N,߰j͑q=0x[[a|)B:*Zp>f! dw7tZ5+ZN[˗RJ?YQSoA#`$DGHU}tC.XH .'s4ӕǵ__8; kQTS}0`b0iVWH8xwm+;'{*K_lm/MZ~\&fiZ,JVY/fWwh_x߿Ix~"pr26^zM~"N(t!;7x|hZK4C2Z=CYJвRh~/ЯxfYT Q&8T_9hUd&D8+_a6t'#/V󽌼YfSosX0zߛAC١W A#yOzFm:*hKc0eFk@v핕qis%>* g3Ce3F5Ͱ-v =ff|ڢ.r*#X!Y&$]zi-xօQh@1[{ȫjVOV> A;T,gn^5 n!![AnRT2x s$/Bz'KFfhX%]n&1 w&ܢL/VL!YH#mz B{Kbڥ&uNjσݮ]d1W9sjy "U)@=Rm(7V+*pMlO(  tsWsȮOy^Ҁh8mR B8d_5IUQ wHM\q5u|rqЂu`xۚiGl^"oEzw^Y7I8HIOJd>GK%C G/2teL5ZH4#-|3@#EHIx +2GޕƑ$翲ЃV1`,`}7'bPw4&Y^H1ۛ9@PJw|1I?M4 o^=PdlɎՌ'7$؟00}ʿܭzg-~??~./˗-xټgh"HϿj ˋ_o/\w*/xn]ur9?>5.ioX?ٌM /+͹\iZnM7c[4DojsiٚԃxȜa۳5 /FfKza=s~FBw:OrgbLon1hhiruMȨ6F W>C녮< KN2ȨF?XwIaUwZt4W4QNk?jxwϘGE1*AYmsU*| }a=E <@@{.(ʱE댖Tf-QKcRzORobXm҂AjW1Z xPi^]e͵큣GZ8;<'PbMxeWkWOŚM4)pU IiQIͼbBByS 4 ڤR11jJ_ ARPԠlpX#sV8/M;)0 AǚJp"*0Σ.9AQN:N!EL'cmz{tUd8SrȬR%mBP? pgsCn:@13 :_;& i-ck7*/wѐ1{Ed#|{.8V#TQ(2x$D 41#H:E~ZQY11Y1jl] "q }Z E GN rR/*J9 G Qwa@!t`9TZLH:6GeLGdInح]pcϣ 2Zt9txQ-R$u5{%.A ްJAo8f pI0ёt9qZ)%zt[;$mPI H(z858tւJ@ ,B Q9 ÔG5>q6m{]4c32XX F- D:,9mrI0m肛 I*I3 7:Ք %C H>=@܄i9#4x\ ѹj"´+s_iN!E֔ 6@16 ͌R@ykpl]hra"33Р5ƴ&*ghƅ" vmʷNU;cآ{$_ζy⊅V-h Ti=Q@K/]ǸJ#y{>ɶp6&"o}glsKRw B_VEL6OVsAQzӥrnu2-E4JonЏ52Ֆ5GƢy-pѫʹ5Hft9:^а`ܼbK?۳E*& LjAPb|t6r_Lr˪B#]}Ꞃ\md֚018S B D,P$3ĂI(^=Apo=A.8m}Ej+$D % n.E~,oS҆ NR7I~/uqyآ^U+Ch !Iק킣)ammVi˧zl$HnFEj}AWp L H*l u/>]2KskAiB­g(3|JPR1M  hS6/k"e>Lz/f1 k<ήt^b`rem삧;b{Y>U˗$rJLG#HSkB19B;@V|b_//){jἊbbAk`YRWڰ#E|ڭ֫ݪtNl#>C|StNs&@{Ӻl5%<ӍaEx>6YKB{8 Q$[zQ(;q<֒ wᑯ2@dZqRR%B6qLtvq0٩8N# m}h5$*ƿBR=~m[db$gzQmώ/Mіn ($oA] &Xwv%d07[3|SzSZ$ )+ۭs $YZ[W nG9;Mφ;vzY{fb4'9.8vO]j|I zL|*=׌ߌu4]vL)/+}.j 88[5Z]f#E[ 8lcSX ?NCMm-<]yiSV5-J&,~U&({jpwa~X\5t%$jD#iwGJ Xz,fkkJUrTv{U]%|. qtMwXx2g:ZZ.D)h ˍ"EfYal7 h76[(mE >8,ik(uQZ9 8 b:(Jآf;g2y`Mʒ][Jr)b`ztc,s:0۩Cvan ǠM/w,v>8Kf74np?Z#U98cKQ^-I#l) ݭ B tAp0©#15nQ#,e-C1}#ߏ[ӭ=p\#K hF] l/D?p6^BWLܯw^Cj7~Ej%XO*Ьez-Nk]9&ՌǘYA%o@is C*~FǏEECtnM_k삣HJejlg$X* &8Bc[#Λ"{Kjwm]C(98jx}&wΣ7-61mcL[ZỲm@d_Ҁ6a8Xy%)m]jp?SPԠl{YG`c #tMڸ*Z-Oy*:-ʿ P\!FD&9u6 OnR47[:K,>.0u?kB3}o,Mh]3C/KSNmy]BD)͔:T֟VwдzQͫe!6GWfcD9&Ι4{70뺡p4=wM` S) g ÷8nMgݢ^q$50QԥvO% ?jKxﺔL yvѢ+I 5Y)ms.<"iȊYsDm=UBA;%ɚP$=do9OS-z5E1+Q(6M<pD;A^[2/0EWM5i}j! j+{xL`ū#YĀ(Xq PJ-@siw: Mcؼ_C15:.'<N򨒯%Ri;1zϱ꼨:j}=)5Բ£gw=x}w)5gjsS>2S&jh}IM( D!v˰ Ί]7С[y ;vO]9šwa5Jxq(ɟ#6|{`+n-<ξ5dQckAdH:Qeu O'k.lDFW]Ʊ[R "_>55d<I=i߽Rg4fQg4=u'B!EKzK7TX_& ؔ"xu=p4i!N\4kE:kť1+cSZTԔ|sգ d?mݡ5>~NB-?3u'9gV_ںtB.BR]"Gե. v͖/SYcܚ07E XXBoBQ$Y~p]Tz6Dq@?(1R㉱N5#< `DV'^ٝޤ.8B ,:gU%e(pƧ2Z(koqկ IxGsAשQڠ*ZB4,IG8nK]p A0lb*\CU 4kmFJ{C3seF&ցnD3q߯ݶ(InIV۞,[MV,V}EIbvy|%# JjQD2κA e*@3m(~aw+-g & Š_y׬j`Śd4ɠ7[7Pe7sYK -~=h=޶3S/I{_ėc axݏ~2{ŋZ=kj df2||h+ɠbN*#@1>q/ Q)QWl,o9=wLM9NVJ÷<1f&5߳oH+Tr^$]5(䛳o;P4ޜ ߬-՟hf;FqurPik._ڇXA1)L\\ ?ޖ`YI4l4dhӣgXDrfI9F`j_^ߜ}.&WYnQ^2 ut`8z-׬5sk5 ^ҭN d[|앷7`o&ٕg0o}Uyݛꦛ p8aϡiIô/r҉N:, 🿫dU["_O̾OwޔPn ֳ|?V/gn=~t5\''C<IŦ h0>>A?W)td/ ~MuF10cEIYt7/;OǛwjv?@a~\dwUS# ?w~6{ukaktK023*o@$I$SЋBKn;J{Ҥ:]ytbRUݰ?4s Dr¼)b_,lyqDA}˚8NGJF~Ô_Gv~TZZPxV[5;Ѳ&}oG`f}˜ w\81#ŦwXB7ֳ4ǎYWymzj>@'Ċbs]NEaWn(^mJ$,\7 >Xԣ_Nq=s+D/ít,Zbkum/rq{k#ҽkRq~kywQYaz,'k_U7C'pN53ƆceIyU? Ա7&bFBͶnv9{_3Rzd۷GYE1@W Ou4j0xlSs??mryNj;Z;ʍĦئxsLYcRh Df,d[Fb>h*.Ż%De!ɇQk1/-qW7%1q@5Uk"¬" &mTۡ"kJڤ)mv)6乍Hl,HaԦnw홍#,[$bZ6z)J2`%mX7g`-HAKm Ҏ}#^GuZf*(Eo*(A]@xh%RHn6iCzk#󓆍dΗc.s mO`yZdӏU+( zތ}(Ɠ7F>TOa'iH\3s:NYIocp:9h$LdzG ~FƇ?-2 0݃ږ`bxX3ɾĝ(oW3T l~YG"e$4,Q-mdeΆuMRi}F1X,5v,sywx;QnJ]C`УFh:uzCkzq+ſ".\_FO*v epx;I +[~\K(hxL4nivGr֖[Z-"櫱׽["˔ĵI8K38vA2c/ڴ5!Rtw6k~ڵ?5M7~s:+LcPLE¥(3?7XۃNQ`j2+T(wrE4ozM6P?;!KTӍ;Y7,ʌXKRc'%>_hYQ||=Av=AJD6cPd;Q'|e<w0K{*ȡn:w[:o4^kT;(2z-XfbJ peU#t |+HցB#]>RC*<^e#(z{Pm<>aul:p .);#ߎ +90B 1BLzA9x1bdh]m(|}0}?11W8Ng:(`{W /~OTD '_L> A"hgFc vb\=komoޟ`Á?,{j7xV8f8&EaJÀDqlPD<$Z&~zgޮ4]wiZe8E揣poh8d+oeʼxOc3$T:MBAz%\K OL4g5Ƴ*y ,B D,OݚS,99n4V+P}bxr0reL̴ =@:Yű|ns B\c&A4(mssw9ǁI2- z0a!L: @qn0 ܖ285tI$ז< [piZ'i2)r,!|>N1*xIsxVRЪ.[Dv^I)NcBH Sd(u{(PŴ -PZm &hܬxR%h=i 4q/t2x4PvKdX.J2Y|(kKnI5mLۄƏseЮ%F59AC8»? @;̈́&O,шb nf#.-_TjʴXVZCku\<9|8pIEjd9ѭ;J&FS!"Ag:\Dц~}:s e9NQBGD3# )9ojy5"|[,VkUV$\ "@N8hL 6Bt`Ã>%Y#k&@[hc#1)b2|zCSl( m F0.պc މ 3p IUG1Bhwgi(fM [շ&E}K`ӷ4]Ξi&օ@|C9U<<"DxNa-DzTk/m4޵6rBXQ]]$nK b 833%*|-W=CR$JRp6kW쮮>,?1s`]\fdsT2(CUBdΙ)KA5w)]QSo4$IΣ%w$'p-6ˀ9H72:䏐>BwQ nŧV`r$g#1T1G U̘Z{7 1H2)6xNR/ޜ"[V@1:dSxV7"so=} HOA8 sibRsܶ%[Hm>H1PtQDpV'nĬ2~M88, B,T{Od P $lkwd#g\`N,)C>:""5 sъgActBI*{ϵ>a[`.5lt˞ \&hB[J]"e-,>Rh98 9䜱FGDA`&wa55.'cu6" ˳ɰ$u=WSFI}+,)!,{ؤ߬F!mFxΐ.IV;@p4H1iPSҠvHVݒto ЖY]'*b^M֫=R[ x‰IeOza 820f1| FnSxft,u{P&ɂN2JPC6'|΍k*;Ro{0>IO! XmI,ۛPdH9]**Q8c󒡥2hgY$LBDMɠR峋\9q tUK%N$gg TY(k$x2t ,lO&2ߚ pv i?o7åC E0J)@XQ}J5ߝnܭ%] e–3`#D&km,Ɇp"$4.4IؐU5);~@24n)]qZ1ihr3Ȱ+P\?z  -2aZzݡXAQ|Y7%@i ̍MlwQp#9`m/e4eW, },ʴ\/Crه×9+p!+AK ܖOrGxHOeguO_PۛN7 [kr֢}amM7[7vDHYv˻"ڢ UYtk[ )Hјp}EXє=~I7Ipj:ܔs4^/~-~Y4MnrUGE+*VZ ?1~_D?/ޣ'{FXMkF?m}*\j!Z}tҷ\IF0#aI~Ol3ƇAMZi|$(BzFoi AsZ5CQ|jH)h^w~__o2˳A7o{ 1GκoάM\H Yw)õIx}CEGizL"`g6 3NOI_k)ԋ3e`wo>===?=Tw@X{AYA20ooE}W־oc?<;nƒyAOf,9 8⌿nl6C,M!tZT(~Nf @C-WjVp`keUoԲ^@>68%*opezIl[ՠNwGC#Ct9'ݷ+>j^_%`XE˓8qzɏËx28y[]w޻GoIb^\.4O0t1Г^:x>} :W˦th1Rw ,uMn>0 †_?}źAiqާ4Y_}kNIǴb՟ӗ GӚa'ddpmӤ$߮"A*_]4KwXh!Q@ja 8ߥ|4͋48Z=g-h >k#;vA^sv޻ux!HckgڛYX"K>z4 @ի_6ӴIל+ R⍔h /2E֞~)]ny*xqӔ2FvHw/WY7} <$G:-l;Bf(&H-lIw5;wb@ڲZnm;ce#oYYݞR0LpO`L3¶յLiU]V:? jwkZ_kr;nq@c6JygfE*{t#h L ZʆEۦ\d;|P#VaR5߲艊/IҔ|DjH&{y2>߽8aw7 ^"]Tr\(zZɣGM Oш M,J>gC9EJOמU+UWF zI='э^"U`DPoJ!7O^Y!wD6Po/˷ _^=P25Z˥ 6?.@h6hkG"4ju6Ji6$~6dUmCoGA?(XK6 _RB tU- [?(WjmHZ?ku]򥴕~rv1}k UcC^XUI! jja%@|&òle^rdꌒ⫲7}fq^שȴ73 a/]o ŴLǣYUmbhi80  PZ9Q4!ժ#}g7#3G<@4G0hM`0!`rfpLjYЊ'f<:F9M3R bTaj%|1Leկ'}ix`4-Qd擛593nKVϾժ_k&ز^j%\zYeiGg%(/{q&Dž&8D.MNI`t" OQAP2+@p('ctysoJloePɅ|@[O]O?Y^)A#9/by W8ųd ['rnKpL03QZ#_o;/$dz BpqɥvG"ta5ʅ}? y9WhQNEfi̹&aт.3ڃ) bߊVXNɥD 3ހM)( 8΀9YtVz=W[k Z L6YY=Gd@ Pa!@VIt1gsY;Ѭd`4xrԚ >\gdFeVi46I.M]$9(d')d%wR *!WND KJVP*fgxtL,c\IpiYFkXV\g@x5iΪdHv™depƑ[';5h@r`nZKɧ˔dȥhIBS*HBl(K+5Ѳ(%}~1`hCYf*1,G :49$׊)eGu 2';2vc"=dۖ,L/=JZYH K`[xL v((DXṔ*Wb;i{a$~T10 '|lU0ͤ&2ƘgFRaL(`:fѥHp:E .;6\2Od!ˈ +eY,$&!l$*Pŭ@)CW ٣gZ{8_!Uad;;A20 cTHJS|4eQ̦mfoݷnݺ:20(E5H2umH|R=^cV Ev H 0vt`I6wCBP8VY@C6@#e V F2FK"# ƆAk,,tM;uAh22K+L tIX7VRr ֠2 0  u!+9dpolM;k J8 seԸƂՔV9gem 0Ac委2.Tˮt@)6{SQvJ""z40M@y!rWóe^PX J!m`p,(qA͹<$ h~ iQaeneƋqYڕF:'WGeS0BNpNyd S\#vfƙxZpo9r7:Ez|D[ޅLˀm됵0pxY##/[Tf%NHN -*``,1\pLN&*%5: .9(|""H&jZ!m2gLtn݁@ /# d anE,Gፀ *p4/=f,XNe~T}^ghyvU;Ug 6M%xYAR#v"}vx=pZqgQ'c26Df <`)VZa9ʘGp;Kf|pj0pCD% AH ՀR"$2GExp*FgAD_;&0*xQE>c9ujP̨< B&--3M /X%I4B.s7Ն#tM3KVǺLd1%$b zV<#Ƃyp;( 65c@8gAM2҉蒐QФ,AΏ 頞;s46`Rpf}ެQ6dTYzKuF ,ˆfN Jw3 acI@e|$MXJ eaHT)m0ZykB'c3`4x;yp310H7fs-s8n C]HC܃n8Мcq@tPET@=\B*сGPG>aK Fzԏz uT T7V"LyR$xrpY%r*A'ynPp^hQp0RȐ*dEW=U4έ~5"hSD?:[19Q ZYTXV=Y$-㹬dü.f0QPQhB߳Z0aln/SRH2$W9F [P7K܂#-K@-o1EA+=/Bu~)Q28 =9&Q`eJH.NIw4 Lv5ZY|*4.X1DTPY4x# !O]Xbǥ1Z`4r*Zj# P VUi3 ٲ, ekWٚJ+om@sYSFv]~C-Wx$Lh}B%'-|p|>nAnhd '.ͨk6qx!ۭ2k-vsv5V!١W*n~Jil(a'6[=rVf&t/8znE'n+Ot17Ztyly^̫_Asp}WÞCvqτ$HD:$!tHC"HD:$!tHC"HD:$!tHC"HD:$!tHC"HD:$!tHC"HD:$!uHXF 4!\y:Op#iHZtO@O}HCb؇>$!}HCb؇>$!}HCb؇>$!}HCb؇>$!}HCb؇>$!}HCb؇>',QH犿e>=ʆ]}2z#k~*zb-_ިX{~ Ho !IoDz#ވF7"HoDz#ވF7"HoDz#ވF7"HoDz#ވF7"HoDz#ވF7"HoDz#ވF7:aOIᇇ|9b{l̀bX+/)؇>},+AAb؇>$!}HCb؇>$!}HCb؇>$!}HCb؇>$!}HCb؇>$!}HCb؇>$!}4zzZ6swjGoۘsZtaty )i, =Jv2$XMF6I,s RLB+*}* XkVt%ÀmYqB kfO&s$b QB@pXWKN(`%O's;kuV:PB " ԧ0o$bЊZfJhfe-X͗%<4W4l5,}5Z\,~n:*ˢ,y ,SZ|=uf[ri1E5lWƗOk]{Ѭ-{ g3ߢ44//G t>_Oa[ aW~+oMh奟;|Nek>3|f R>Jqd HiRR2|5Z{ɯ%l7o}|$-z4ze^g7e tjGG'|v5~sqggvlRh.@pŏ?gџh>! X_v?~R.t}_R߾yG M_&l;gUA&J媋F_/TʆV":bټj] Y'(ۂ1ZVv|"G+bv3M/a-Ms;L>E4sAW˱OԮC3 ޘooB/Xs ~1QI RfF7klF1L3*Wz7_p(Cy%E*gܫrty/UR'wg`иT0Lߤȥ4"WK#?u܃iͽLx#m3Ue9 L6 ٶ($&ۃc3 aC_ݱY_z/۫l˃!a.zug;>j姊훋na{|~|-G?q9sTAZdj BJ2+HTmS^CR)gu3J0)mgHyM|F-#b~%>^7 rZamp8VPdۂ׽#W"ځZcmfXsjϳIyez&b%zՂ*j=U1jr)7?llnw-/($p^vH{wW/r3'ŧlֺSS+>QKiu oDDA'?y&csQNր25<{CV\-&:vI'}5<n8Nӣ pZ[0[mϫ[i[Pۦ磾,$&Ƌz%Z+q^ѿookH| *ƪsekZ!I1oEh[-m}88ԑVð.ZOlQqt@jiJV;sA[UAbdwlwukbG 2"RSчSe}YnZ˹нRCIkT<ޜ:tujx]ݪ(Vķ?_o ݛqڜ+L9W6eaVR֪ *hdy2`~\J$_uLpbGuzetWPrՕc.i/B;C;gϫ [_δyRJ: R9Z2W\OFO.xe|1m74v%_qM8 HQweY` /^6fdTWtLUIn}R*H)R̡`[FT~ F,ERF&s@cƑڙj@?IB_/v^f K "#Zxlj cG.wYk p3_(3;dwgI-IkRfwfz㡄+qs6lysv`xK8$WR&siE>;ֺ'x'`тi(_o0$r @:NFdB1G>ra}_ rŹ4l4B[RZ0ό!<&;tt$]Ht!HJ\jE>][72$,9NFQK0yV-mFGz!xb[l*HEdDplǹ#'b0n".n7&1IPDoIpGRȈ&A$(sVOww~bt!a揁B)T u1=.kNŒY=G>f\p#[4(S8RJz'6KP 4cx Ǒυ3c\or(y˦9O;&XRk#Ɩ~”e^hr!(+ՙ풷p}Gbbb6 m툌@=s^J%YI}'M09l_A3Μ#!(b8ET_ %sW/e˯l v D WhL\ Usm5ubL$!`2M~6?DFFf5 Y%mMXΨܹ@ f6E>TW[PTsQMIƁ4M,͌ːcx Ǒτz9 M6/Iƹ܊7|(A΄ [>r|yMHVlG<^ GhŴh 7<=|䚲zzF鎊$d;I<YO\J!Wd1 "+]yf2xV"}N]!idD+qs!uD,:(߅nX&y.JxgH"H0ΗnEo_,# RM\n7jZvnԌ TvLu"(0;SEC#E>Ig +]6(i)WNQ NIJP{$Βn%cOYCgBramuB-&uDFZzxdꄌ䵱\Â; DNp{S}b0 E>dO(Du_{ep5=L) Gxbn!x-3 {T~o[~y-ԇ7׿Q1Q;П('(bjHGd4\LAVlXmF|zI@xY:^Ckl4w4* r2wmL%ɊAbYT^%cf"&f"(وB+9yp&cg2ҢDߍfw@PM2 OCL }f@0Xĉat7qE:(dH(a :ˎK샌>}c" eLiGAƳiTr6/P-mū6g q!S8 sT2w:HGb-\Mn^=5|DCq#V2^61mhrΪ~" /|. V }oQGd4 ^;\ʇ[PrJ=Ie<-':;TRa`H GS@/}"U#Eױ]*rgÖm(I%-n#dhŞk1iBm{քz |6L$uDFÁTC+*߷:]~.oˏa*NG'}LJ۠w|*e\F)0W<?SorGFMtC 8<ӍPh+,M=}g`qЗ{О)6Y*MsVdLț1#;C>t`mԿoRsLiIta`m[A+,;O9.~L+0" FWlZH6YBJ>]Gڧ}ma7khiʘbŒ^ӄET 72cg`-&xv; Mj%ꂆjq.FT,Ql~M.%` c(#I-4 aAI9oh*n-&[YϚ~N>pwPxT.-hX[s[7gGԈ38>! n>MT&ܝi9 R?0L n ,o"zn]Bn-|.W74.ahl8#jl0w*Ԙk\/g@ծ 6\bi&bS- T7`AK m) Dw9٫T:Kq‹>pƽ$j6Q .oX&ga]`[n0q穀,ec"lX>hٳ,wvB {DFeVP~QtYeN~jG: ~뮟'K赣 rՔ*РeqeEtGF!(zDrMIO;=!GB" Zxf~LPh\0r&! @*t^r/CE>/Xz5sF@TƊ 2jԱy jg:N0ޅU/|  :5-#2V,u/ͦ)̕bi&**ߍ>v$Y+b%Z,4qIF^ Ӿ$ٝ ~(h! .h}?%ȉ~k˲Z!Z墉]- Geh8jT$XzT 6l3 *EfI*Gh3}VM :mDF-I:tN:]^q x fP3!7t8Jӄ0Еp$X8 NfƸ7392a |.b}r4y2۲]b*Nc}\a&'*v/hl!]]ߕs&}G(qĶe=a=qFcI')$(\IK;\D," jʼn9|OpμRsknfNP[T..`>5l.'f}68V'wSMͱ. h?T7`hQ\͏)mˈ]G]0Q9ʣѰhI;+&\G]= poi\7<| ɒ ;&Tq 4 4J|d:0pKTSt]Y2 碩lS7#c䦻Aۢij׌86MSQ {"A3E>pjnfUUMpb&]G*:LҠE@_3<ֺd-jE" )nz}}A٨w t̢v#a6>5SplVz2G|zep'[fsy^j<{<-ȊS 2I J9*$tVw)aD E>{\̈́ MC^q)me~ώɡpJSThy^-S5KVŀvk5h ӟE4<)[ڹ\7ũs0)\ȳ i9-ߕFNGx}]*7zBⓄs]'һLrt\ BMAƛW1#C>rZP2Fdli۲wjB;3 ϟrf8mZ[,Ƞ1?48FTO.Q>OS>YZ65y|ID͝##2a|&#ǚ󷗏m}u97Sp;b`%cH#,up xrdLBߪdتE>RIԘ]WZq-[;W>J}OcYhqQsA8(p *k ס=\j< NR` v,18+yKxDF<E> r ^k$|n){BB;0o][o#+`_87npgqvr$&Xni-{f[T-QEY KU?uaj} S| *F*z*LJwR*j~TZX\Ffe KW0fІC=CY%'{]gFțc[ G[0'jrtqc:&P(U$WpXQ*\sP3aeʻ}U.yM~݁d3x7ӪF3S\Y)LƝ@NEqpr[)"+6jm7n#x}C`9EYy7 g±Ό1@žšh }Ys>v7J ^B1pv<ӡ{dzM6^G!'ѧR3ik-dq=.q3Ax{3DWIJ*dPVg=!1Jx(QT#P&ξ9h&T"KQIn=%h1΍W0%iWY.BV'y{3h),>)b(Ha&U1fap,PqdlK?뢛*#j>< /EMUe^P8AIBO1n `W4v^W/RNG>Z[/҉ p!pe\h0A +}52^]a澚N׳Wy9 ONzLoN Jo'|):Աv9Sl *Q DLp&OS}9JCfQߴAcx򀌷[Ƙװb Q x*bd~ē)+E(6HW Μ)\Nh@v`U~`׾F&g1f.jXP"jpxSFetɄLHy =-v2;"\Q‌K^&T1}}0Cp}BS".SOE{&GAp F3 Jý&p4 ڮy炉Pz>t1D{@S_0TSvס&yR&$>vۛz1Po):2jGx &wi/kz\ *,2)z%\W"U5]"*p ]&Ez FSkqpֽkm^ƃyҜ")d1%m[p<<·V?_=gHY3҄{,ECK,լY]'=%vkxY. A3'f3Yweq9qxrV~; 1zyy8n%sʰÈgJ#fsaM y.>$ij˜2Jv35%G ! ^S3(C=O<)`2~%GZ1^ @!Q,0pdtr$:;Sk?3d0#f`ܨ3k%\h]G`^X!Yx a-~D@_φӿTA}?w~tc4D\R_f~-Oh &~^5Rv?wC$KCHK)Ť~禬y|aD 02jO3ʯx,bL&=a/y`.G3m-i3e=׳y97oWG󗟖|w_| gfqU4 |Jq^_{uc+ȝc.fh)SdK9.0 r0n=da~k?+ Al~)oCم?lќq, iV",wUIeMt%lW'^iDoWÊieA[6Znʗo=tΖ]f~5_W=DIijۤ\@Ws+EXIa=_j|9n\p^qnuq;]Wݦ-WQ_ϣⅎ|B|BK3 f?0wFheDe~=՚eiUU03В+o5yE Ϊȁ>1釰Kp&"EqŔu3}Q' fv;"GbT͊ NѸr ѷU Yͦaǿ-any-dlڧ?Wmtk]@ 3k iοag{!*,)Px 1L9֎p\- <b!i.x䖎YBKGZ)IʥfyQ#Hnwm7˖H=ћn108v58Y*K'*qbUT viZbx"+<.&tC|p\E<S/x*w^wyB)Px$0Y ˚\j_UEI"$ll!ж<ԄtEEݶEy}?p u([0΂4R:bP3p,\2x43ˠ6CQ 1m(!c\Stb4H'KDއzlcw+Z|vш6+%!odccmqDyP-06ș I(1Y8.LJ1F^C)UMxíAK%Rpt9o¡Ĕ; "yc-1#X#E< _SpQ*y3>_}bޝk|֔ urögaT 3"GS _ ν62sVkt?XI32 F@h{}!11:=i97x-FMYu;>PJшɸya.漆 N3H _{@R /f d|na1xa,]cBP%UJeH;VAA(78f<65~1ıf3J%DT /6x7l?v˯YRyG>8d93 piu5n /NP( H(Ŕ;*۱o,zL<3`S%+],EEeiW'io?x"Ez Jg99k&L#F*iQu\zN 5^Qį/='tS,43]Q.iPM2i$9C<V:G2P6oWAy@ q *X`:!MbYQ# 2RXRAs1hWF Ch3j׈X,90In8v8R#FphW Y:. P`ߍ6]8~ aTڭ(\~26_ qQyC`(a[_^CܩY%=؞:ސi?9W:wtJ9t!2aֹ3 lsiVE\h0~' T؝ ,;OifC׮U]: )! "{-ӧpWJ2*3f seF[bQ "bi}iCs>lhTLO-6+/a],go(4ڴm2e}@54kwwGEWph1<ԎqNrr#!Th2gL=_TktHVW7W1Cpud<$^/:ċe<9 Eg$pG!~L,Ӗ'DHrΚ)Ai+vgHa}; OXi! TlљRPٺkS7t:otAGHx1{^m@rd쿘\U&%N;&8>5 ?y L[7W]RNGۼ!29"[\TvfT5>+sCFtTGhd 3}^ SDxKI1h4[qWƵ gow_|_ooʆzeDۘ01ıGe3s"wkqF8`H-;lȇMdevێL0%Kr۲MP#c*K,>Ȫ_yD`iN3PRCfs?K5^jcU+(&e=dȐ@ pi.O q w =B#lCr_ra9I/rh-Bc%x kB+) Osc81g &Qs p.WѫhCnXQVQxlW˙q\Ԟ_H!:[Aiү 6Lߝ#:9cLؑ":YiZF&UFGV8f dMbX1bY `{y >QF,ri$T/("ǹ;^#6-HaD{=Ha qtc:j*Ţsd:nQ3ėp% nDexI8v)Y11؅ef1n " bak1FB$<ʉ Y;(']xӎg,e>䊺pp`@3贌c0ajbPU^UuM';'%(+vj0&7Z?^^mZbMM9,lX&{9\\]qQ_i :_[`q:5!ʡnĖ(~5c9 g2)1[ jTNcJW n $T?^+_N]@OU/' Wt&emv.aTq5"&yjr_!&~?e%Cqn=F'T\}NT[F7٬?zu`?YqҚe2hm*l ĽP;9?̪0)?^yu4#Z⽧CR*=+_^ i؛GF{h *F!VhM){5E\[1/)&@KNRAxR*9Gy v:Zl0^9A|m 0D_#eՕu< x7,!vJe,Y\r<=o;L)oE/t)ӹYw-c,L}~m,s¼Gp][]G Qȭ9ziuQ)څ2oPY(kf1lj1%ab"sHnjr+(C|Q6<\;f|%VXJgc=|O9%9J5v_'D[_io t ΢~hQ3MVM1'yBꬦwB}+v #N5q` &oC|o#e°2'X+:RA`f#/tW{4cTd fxԅB^P-WϑSMsw蕶n'kD{GljcWF+NsMzVafOzC n@E;ZR L 9b S`a }g=rjv4vr͓RoTfbgcT&T@Έ?yso7m'040F#x\}ži""pj@+O_hSE0 yZ5ur a1:)K[. ~Ý ˬW& t AK6"f1ʘ)A#P;az ur a1F,iLiL`|hdɂg*ǽQVz(͍o(90jj1yXkZqk;pV<]LjEqZݶ\͛P|]0뿄+]@ E ,LqAW _φsJ_}=`t|*h>]觚z]/߿^XUj14Y6904#{طp anmƃ/43eja}a0}>z2Q<|vo~gǿ̬Ó; zGa}}a)'&-Op=^4b>0Z]T݀9QEOCK͓kf~sJt_@@VrxpK.a_U)o Zzo *SVa ;ek5fcRѮeicSX iާvt5l(m7{"t׏WU3q6˾`_S+xR~lmLژT+XWi#n$n f<]-V\>`b2מ+ts0_Xwx֋eU^ۥgB\;_Ҭ|;U^+`2 F@]F aНӎf?GMެeϣGq1p>>L5#]=Y@Ϛ5Lv ur|>9-@A~[cC}0f 4)&PsHY^m]ྶɞiɽ}y2]v~krD\b5P=~ q5ѩ=7E럁{훁\gTpf\5A|~&83=Y(Oꗸ<Խ +Eeq=zVղD^f/֒+ ge`>r:܉ͽ?/x?ANuo,}sGg k>Dm/` 6JC&4KY€T!'5  i- F%ΈT@6<3$kp$癇B>Y`CDZFW3]-u G<*>?#gfǷ`TLQ}>QC܇A)!ߣJ#Z@5fv)kTI钉F4\WNDG܇PWc`~qr?Î=>_6+< 1Nu+P&(Z0L~!8*3n|4]7h֪ݩ.xx,a&^`-D0xvK! )tYNb8vj"> Of8HωOo(0&#}0%]ʕm\)(OڡxARla VY *74-g3x/Fw(I;! ̮Ll]38w|X䃀b*PxXΈX3!T:aEyY٬-f3Ά_`/Z3![.`(JS D,V#'Q?O4?7|H,$*a c@(pRBԻ2 =acݶ;,s!obi6ruO, XTv0ξ钗q fgV/%ވ Ol &G81 p4G0JsJ(9^c5ugOooÞ P2w$X2! Ѵ%ׇb1GEF/ESm_n=G0Ub1Ge3A #`Q ++{S\j H0⍢<.S~1`оࠕ; 6WnµPTP<tnajFhF|@"":KTrxĸ%H^-k9!pLǢ=mCS8l*r#`PaFLh,6I2pi-ڨ>xS#cDVԖ:yR"BEnbĉ]L_nT<&jMFxEGxGcbqjdJd騘>>"?}tFc!% "1*lHtʸD3Ya0t8v^ vL86!.N{d, #]Ȃqnu83wv7*r2)ΒDgǐƐhϚ UP03N sCT9C6hCn!"!bc FQ F d;0ko؃G wc\hy ށ!>?\_%`,pf9'W9FPNhFm/HL"zm `zGZbH#QaVMn|5id(\U.XgnU]O6\//&E5iQ?/.}E_yqrfcBTsԒt{к_fUwktܭww2+fkRx#?un,;ϋߚ.eV0 QJmȁH@$#I43W:$/noIj ~?0~:W}gH p|QO~5 / "cMC ϲz^Ovۻ+o_|ޟ$F4(3㛝ASe:-܎]|&+J띒UAD ua1F(;:)r˸A!RTdBبɓӝBhvTJr'Di(KStP裞+i*zVX#+L0.2t譌r+mT11Q):QBeNz5Jw?n_Kaw9~ |glY4a8#v4 aKT"U⯼V+OP97W9tWWnj`o/+0N:3XY`ќ'BP ȧ !:Cm7`klq-e4>Xxy{n+Ct64馋K;enSBqAގW{%]^zΡ \(ש^4[vouC̥(DB;{|ڳ[@};׌xdEBa^cR6y MNCe7>(ʑ3 .uޥ{/q%PߓMA`I0cV!v )Ī*-irqPv >6.UN3~ᢅԦGD3ko󁩛ֽG5v%$RWo%ኦ ` MzT8&*0v?`K2bxyk AyL")|t]l]>I4 <{E$"wGzzx4w=f 8zi H#cEJ-QSͿbX~{j"mEx׮Q@p.{h GVep8: `Bc[-֑U@ @Ni _Frl53,mPZh6MYx^TP*Whۣ%S 1Bvt >:%>-}(1Qf]ExY6Ava)qw=&AJ'ҵ]` T +iP͠Pb*Ӂk%(ʁ?G?%`QQ.gh?_AWMW?'UՍO ܟ};i O;u~7 &잿6V2['] N,y\fG&{i |Rw{Y$5!?/A D]w5[ w(hP|w3w_QCB]vmuM:_hqw+.0Gжv7ܤVn*kDţϨ~z>n¿|zzNXhDװ)3q$z1 񘻠 6K*~ GT:X@'v?_o`amu3Öۃump@xSG/q2Z@d~qGIoLodܜ9x8**=jezѱyY*:Zr^S`_6R(VgU :VgP(VJ[JM>'oS(2yE1,ʞX1JSVUrj@,M7_ Df0y 2 s^)Xj9V8FJa CDƢfP(1y-<+CϡPbo'A m=*4y5IQZ0 HȞϡPd`uUʫ]"Qex(҃Ǘk6 Ѧhe{lxV6mt^_-\FnG6Afb6 !ʌƥ* J4ڦڥ!@VJ <#9 X>/Jڪ`9 `1 i81R#&pb6R;$LU`bbK _(Dbf6[MW@ =mrl7T(jJWqtCi^VkjBΉ q9oL>B+, g:sij6vi&D+14℁#^T.ۨذaB%rd,+6CU6ʓhPW\:ihYzJEt084IN#>6g*}sPHիT=^D*1yFF=^h;Fs?TsXhָq2N j v03YL}b@ZgT􇝈 'SS(1yxm=8 %&O*ݖxzh9JLQ^:Fq$(le*D6r*86 %&^ry^)CϡPdVxF:ôH0h" P.l8s(Pz|PqA1b8#ϠPb-&TW 7E׊+s0ȲϡPbґRH]h`QI[slL^|CϡPbP `s+IVPqnL>Bɛ)N)"C(.^*Uu "Bs; 06*~w|p|;+υA'D*r yAN8**:s~Uّ #L6J3(v)5PD tU+47\|f [Uer"W+Hu7Ē+ M3=za.N8nKcl^`f}ɇIJC4G*@>QE%!ݬ%^iU4Mx}2$n|:3$~xCv'Nmҧqv,5Tx7w{ncH [/ĖK˅ ;;gp\=s{Dps3ȶϻ:hih[\/#qO hdJ~U ~ ,Yʪt%?䳜6+e:L/Ȏ/vTPfP.L% r}o"_ Y>.gԜv8\`*%;K㾟YŽV:JRRFwBآ~B0ˀUWE&b>j>.BH"\[(c◢ U疐c׫pN;T2 !{1, #rpx ;8@lj@sB;evq6"r^-rQ[$\ǩb*OAsԩq'w-rp"=H ¹]Y%ҮMF1O-6f==E ?#Yِ{H~W<|[]^S{/"K6{x2 Fp٩ORYO=P\~مtget.δ. xX)k.Ζ^ك|*;\bˇlZRZo y[@@NY,AŬ}G#XH.ZO6Sg!Z+WrݓdY,@:[_xs,k1K4lCB`vaSCkX?sY>"6ލ.%匶sj/v΁Y '+A:d|K}G2Z! a :A;Cp[.gQJ}K+nRd(Q@Co9]+,>R{0_v\y؃/S\K YoOYv~+wMHٵ%x6 Cbƻp"L$tizBh?|z;JRC:0 /0<<4{$$Ym>A(3kZn"CG /1[% !@- j3Z2u uwQ6]lbBM/秣ԭUfÛO,Q@+>,Rp xS$NˑI/a(yC2T`"Figyǟތ~ͦw N}PKU].ӃutW3.|,XD#l7+d?Eqgue}7㬗v 1Zdvu7R40V:f;T, P kٻF$Wƾ,> X; OEU%7$"3E[d1cddI$WU\amwQEIqlg )k.mw\=K _P>5\- I/AB$IpE=%+PC>Z7ZfTKe틦B@׾*A?*}}WxI H`mw" +"'C6mǟ~{wFgO'K˛ߕ)>0 OCct(}8%FxQ&k9~\+Z]f?ˬDBİ9 k+8L&2v3tw9EnLG*F8a1"E(RBΉ5'`>_*Z,:\L 8X5铡 ae$\b a&LN& L&/?~>_n!)8qv((E;ut2!G ̌!3VXJy v21&ᑪL1e6Ӓ+8>L" {aF#<L үG=b-q8m(6$#X"!}LpIw qicnˀY h?sִ?r+ϫ,lX"<r2kM!"u-ӀQm%Joj%]S. kqȤl;j`.0b#ʡad~NJalΔ0JF7DYګ) `Sجp Ř'BhǗE[TE'>e-*duE"3EaOxNh(**}A0#jՆnD`wSkxcâzRm[ #s\X`1lR$7Y `k>˦1J z@fPE+[ϹT oL0(p`C1^ќz#<SPzRuM5$-zXW`p=Ĵ. p?e<Ήj<nSR8!U[}3]0_g<S =Zzjayui݄%ѬxbIMdz̪ ==n`^wMY~wU#IEmPD" 5!G =?+]kruu.XK#<ScΡ71ء[c0%`ӰƃŴY1,qӈصEMT1tƏWfB)AQw5!G8rIKG|/sDץ{U8՘)\s :+gv t m˅_Lii)ڜ_o GJLJIO%g#w"0A܎h x\JV$R {T^ZMGR糽B&qNnV[T7V-* V5iȢ1b8%,03 y] mrSa.o GZg*kӲvAh`D?'#Uhh Nz,ɽh N[Cg|-o1h KrŪriArvU .%|Ic>oQY:e3]MQh= *WEFx4Gܜ}+Fx4ǜ=h׬APPu<5ncH{.MaBXQ;l4!G]"^+)h aP[yh+Q1}YOLAYuWfGkpRo 82 1#8<>FYne*˜mVr=_\RQy4Ƨ˜crE묹LeP4f.䜜sGFhl=Qi3‹N2p)*WAIDiAzÄS&ah 0- b35x>dbgm3[QU_D`{OH< UՓðe7B a"i= +?ۼǢY=J)yUԶ<_LV?ZڭhIj0RPd~Sc49[{V)<3gIC0ޮlFlХu8޵恡m!gQqDsyn?epY#<hՌS%.nzͧ1NJ+ )-t'OuҬTr9g9(g7gJizrkkneHwco' it+XGpQ pr sј5Lwg(Oź\oY=ӏ߿+i$wyo]Y_rpQ & (TyxsϮgU.5<}C_]pzt{/w "w6-Y]-V=Zurӿ]x=1L#߶>6;w#D|04lR)rkf+M~Q~o}soA vwu{a6|u`Tv] ?].>0vɖ؇>{/Y;-|}7hin⯋rGvvoy~ّGKd˟첏ON JA.O|/D+PX_U4naF@/@| E17=,ۯWo4 x߭,`oUe<ȼ\]_1y[<'lC~D;!?D# r7ז (EkFbE悷y$4eT",bY#TʧO?>G76=-!;xK{N݁%?wn\?I~}muƋ_J_쬂LFHl-nJ˗6'kkS &^-w #*/uO^ήOC%RK'V>gh,Ǒce(6p2>YI_9#w Ҋ2 ҉אm_Z|f>l֋1|Y ~u % /MOZ1o5U !aa^frh1蔸U1Hp,!"7^P_etGن~->rDd)Qw \~:i8a6m9GzI%wEwm_Sy ^X,, 5q*4sOzU(Iq[iapյ[Ju ')E!0'JXC=>:^MIk⩭rF4H_?x:taB" 65פJ]~ 6IILgsS<ņ%c)br!)a~ È)fyrJ4Y3ʴm̦X6x SKKS\}~}(x&;{B3ϔ@ʪԑ,YBexk 8[F3Gѩ }pfZ&Wq#LBۺ$Kg(9탴OJ\eJ AEFy(]Ɗax&ew egjqQEKJ!Ztn MRf<\.-V=z  F7eg$pkCڗ/ {wjz;(9@Wjti'Pi v:r s+LJj]d ~1a6 >NBݸ&=?"u`udlvTYn]^Zh%q)RP%dF %vH{aN<ǘN7Y' Zủ3d45;Aw'ˀ_:Hpʔ)At[|&\v fy6,x12)N'A'2Z;RT挦F뽝T9xN/="0 x@NQɨBfeH!O)ߌLXb 5^p+FuXTz2)&CՠYXW>A$/m:]Vg[8``hh F&4n&S(U$69{$_|p%u꫹n82\G۵ |4G u l Y1cTMbFK)g +-^RT*!ǴMljֳ(z6E-{[|=0µ KyOϽR,ƷG^ץ!' p=2՚h" S[] ֏ sƂf݇od;)ztbxg/ripvy,2YM(%! #-R^+IdB-;N<̲/~a^W[mjT][ݫy'7vʁz ۏމ`4MvP:Jwbt k(V.Vj"KSBA׼ݞ c~]6i;/g]Hp nFm )=|<=,˙b,a.9/rLbuao~|g9ͥRBN<4,*o 4ylT2ݶ~\1|7tlv1wσĢymB@2"YxH<7I`n;@@6g=X(өVlLSy U&:]HN倛0!L"gL>S=AI!X93c<Ǿ(gN׬A׉>ƙ^?_ vqX;cKr0]-F,FqGiu LS e2whƼ):D}JM&SʹL=fXuDfg8N~]!XSӜI a*s.2nTI䈖Z2=b+gDJo#XfI1ǝj+jۤg6lϋ}XVj-`#kGb#%@\Gp}Yp)x R/ȥ.ٳ)J狢}fDk- qiXeV!s/T}?b(X'`| 7"[#P,4HoփSM8IA7T$IPE4BYkܟWm©l~eRhM(W(N S+1z5fLA$S:(Q^;(1jiw9JL6a{yN%=|ꃔ MMX61)c7.“\1G ,NTsʨ`.'G4tLIX2 {vb l;F\FH6g([^`Wq ~_(r+Pc|YVA UF'|\tB+vfd"5>#.˙v$B&FfiII@Bd1Oʉ/vRoR%Fh< (:؋$L[-%:Y+䋻/VC;4%$׈zi ߬S ۡ4DŽCK<τ邏ךfsu3NjWS9nW[a,.a-g,^jl{rg9D!Ĩ ڹ(];JRr}Xt[zc>[aevWR5 4*VO>*35&8séXZ̄JG% F=Ig׷_\z{PlXlsF0xEfܥ:,YCvP4 ݄1D*h2ǁ U&ݻ~DzL&Xt^҅kĺ,Y&&- Rwf1)Q{WUoU ʠUxگvTTMEEÀח3j^R{]1WkFp7l9lܷ)At{X֧gx0P>!(ӾEy|;vu=n*2.RI5*+If#>4hq`s?8HN?6Waa~tW+ulFLB\O>0ɠT?{WƑ R݇c3`&Ag`w NE*lʱwW}P$&))rlu;~}9q^P*?͇@>_3 4}/2XOkbm?UW:p8 {%aV"L(r1L0D431HV89G>}WP.^:y=Xu Zҡ3oŘL4w:SoU E浓Êfk88Ý)ޥ[/W]U+,^;-Ov>T톮+M}eȾکqivٍtsHĕOaׇ&ouVn1q|iqo]zoFo\0WY4U9:б7z3"-Q3LM$9XM3DbdH3BySaq]|nQ+DE#\N"E!)SfJYCy y 2}ZX?x{FI۫qwf ܓUy|䳇y o505ZwsR6=U`j-h$FHB -=8rFsLx G⋰~G (WK E Oc*;CY DtВGo.x fyohfm ɽr`x|}>$dh/;lvXs^mW9kZ-nAκM< T "UJ|03 6#,`Tj}-CA); #ޠ )(ԝ<vxwOrOPgջ_%P0 h;:mqkrTat16;ط~0V筟f{ϱW4E@UݘB-M[_! Ϧ7g2D ۪ Ypia!-m~ =+2w0ullkq`sCzn>L0\/Q..˽~o玪ziqo0zj S8Gn43O8G`ZFB>kx47:sp=Yld^dӨMkZ{JMXy$ ́|6mzKb ޸*Jt7l0*_G޽]m>&?~x z؁ip)q r)a1hkh*кUWŸq`ٸI=p3[0z?~y;Kw7)nKwZA}I\|i_ҮbV[GQSgr) 'Ӣʮ!%0"+rkAE("yPh L4bE$Kyge3bHHtd\ȑTG$5z+)u",vXu24UIy fF 6:B[Kت w3eZ^{/yQ2liɍl$V {ˀZ; 9[%K9[%ܼt)0XI˓1O2=b0̦h"ȼ sr|CdMN.Z|| \7BjsbF4K`EJ$W&&e! 1`a}NXJ0-3J͘YJkw1ׅ7iGâ&Z\s#V4l;Dr@yI|sQ7I<,0cXrS݁d3,f)f6DtбXcB)7j5M5pJɿS?Q i/+^ V> ,U@▱م-X<Y,f!Ynu '1beP%TߦcMLryA*$<3cĜE NVF\#K4iCmf +Sp &0?O`?Oibґj5ZG%'z aqe1{A ż'T*[uRp:q:'Zqg7ar_.nFmpPX [G2c0< +d»pO\x-.V JR5ywo.a+ hKC!,՘O`NؾP6`Tq"7 iim8ї! 6H**[Oq8%G#%1H1YyBh[8;;ev ͊(ZNtm1=c6T F3b:{w^J}/]UY'i~;z|^L,c05%a4UcQmQ'}tZ998>o:@Yԫ- 40w j $uCl>!|HCŋu{x$^vql8Ht[Ӈ*h<^@z/O?jp?26s_@g8 gU29on}tG4^M&/;L9Det|s|%T8 yS3KI VDpCK,s$k%FR,3|#LpLɩ8tא.XMb۬`͢F37yǗ×8sgv11eqh_0cuLLq0Z^TL $jC$UiRRm\0s,Nnm8|-B6`k U cJ#68R$#?g)Oc.*581VٮO=m;1vzr95y*@GS{O#vd?Վ׎t"$ri#|8ohY#yҼJy1';iwh@]:N%_puO/lI⟃O]+iR'?ߜ'e7'?“&x z<;i\}?A]RMNn6[gjqnކ SEй/H~a:3_R<9/ɷ `},(*,\5]DmK|Qnl[޳$W|# 3{$asA?mOl#y_5)JLJLْMluWWWu=*Il&zZ>*>}de/goHbP P-T6}[ִܱ^_ʳۉ)QӘAT;hםd@Wn砂XBu|ݗ0۟6>?Mn9 `hn)1H-hi핺 =X@0>fr  Zv82_=VN(Z\$h?]4zs)^Wʐ#%x6"|lw3RQeMsXh-yF9 tP[/;7.rH/q9~ٴ3B%+چ%(SK,Y̙C:<r+7T´▱(jbNłxgAUbzĂ$S,e2`[gt y.SiŴG`%Or geW }%t/"LQ ХXN:N޶EfI/I |ZP (L+{uv9!$4 |t7~wRwj<~&ܼN7gLW,މáx){|s"!z j M4ϕi~I8yV+ֳ@ ߩq;ϱ&sL)Iм~s.1hP'YRM.XNIdX㟢Z=\| E9Êbt& `Ժ [.|;T} lשAߕ],>Oq&j7t=ݻȵi.tjz`op}yZe׻ETbNFֻ];=/¡r==L5Uo7wĞ7tMVO â6dOo"jU/7u7yd5| l8i4 dAmyONW϶ogO*9y*9yh`@S09i3-CI1XV,0gY mβ|nnKoDe=1VYE)`MleF͝QFGR :K}= =6Z;Z|6Q:^I*`35L)=G#UXb$(C3TLqDzwU`J yJ y=%8OS1:=?3sL(QmGs+:2,"1rbrX+?B"{IAyOA 5<6_v9@eP]uc0Qu>!%{i!4t>.)1V}6%Gtʉ.gDZZN#lRz6V̕Q e5Uwň<3fsKA 94Fy |U^\r-P!hqu{v8GQ:w-s*6U`A' Q,BL )v`l9-RbٱU6k_^$R(mrsbĉDbt'N=q<;%H}ёZ_[o_Uj-tJ]JC!e:{ZAg3Zj [:ofoFv3u^dZʄ2a^!X&(%hnDX.ܳnɺA&H.\\Omif%kbMbjsEi˲,,;ﲲU2P FB*PgI)RihA )0Qt G!בDSFҚDš8Ni%7c̣p"xbFi~ٛ˯PRqKySqfטh1Vvƅ^f 7ٿʯS*UL1z ?O矪u]qwN;/S֎4df䁞>lJFK.)|0K/ۋK_(brR c'X9]qD0ktr&3+{dzsu.u>{WTÀP& -,Iń-_oC}HF5,qg09+b~ _kSleEbپWV!{yzsu}ZS('Jh7!NM|j-9];V)s=j&gKs's o~*^9=x]W%90W(^^zۚ*G.@~s^ fUGbyaH0uYN,oi}i\orlG~|O?o^}o70Q^7?Myiyoq/Ң;pg迾eE_CxT.C>g=Cxq)s-e[EaR_-/_^S }=NeANt*bb 3_}6*/9^L37뢀/+Pq$UXt4#YgeR9Q1"[95ӊ "6<DqtF/bƀ.*Wȹ]M~#b R)3mP&qx!~I'S.6K|Јb{P0v`-x Y@SJx-`^1mϒelM7-WyBĥBtؐC:<hϰyP-'Yh>j-}6FGϣ;(5!-[ ClT(Zc}1ЫվzWVCՃNZUgeY:AH ,uLك%蠦=)5m:~y *ܠuVn;>2u j^r6(C5A5e&EqȣطXGvo.Fq]AطVD \j \XzA]7I^Mss&>In3(E`C6uDnC 5JDk1`m4V Ͱ鬝t>uxޏkS KA6qߢ}$Wzf=L P*] S%w3cPLLryå O=cF@1gz4‚SZ?7 egr%KQ~RumsQМ3JX&)K@h>Q)_"HgX.99&(^9/1r=@ΰ<< {fj :" "=w; F0߈DiJJpp^T,ZA۠FHh$0Yfv5L}9"1O{ Nrg8Z`=bҏ&-DisGI(n?/+m֡hTy MqRjpE}KYA(qnZ8<=C"Ґ[ (B@>h_"%Rc6XP|fP4TJF9`Ť%B"D)59bHETB@Ѵg> |vx'*>T-CaHdW=3_U+{0# !ēhJ.X,˨wTc"=c2C(dG;t֮o\©]YE|)P,t[0i „R∂9`Ms&p8` ̙_9M0'ޯ5O%qD.ND\O{u^EHM#uH."6!On/`n1],I E$LfEV#kS= 0jz.&bhh`n+%bnnGYqx%21Jw>BTRX1c2TDk45[.i+ұ{+G)69'N8p<%CRTE ר]QQ8RI$*Ű[)r}qH6Ȝ1ע~fyX W[ZHje}G˲nmے 2=-QGdZ})fo lZw_v[7wŮ'~4Y(mx\w_:ÚFc޳rX#,Rl=.EY:-. ;[q]k9闲bf{%pV${)ol@A26`AE4>ș ^&h鵌h2E m&iu)` V;PBXVڸQ0)H eH 'AM9,Pds] KX˜E`LIjR,k"s>L:i0"ari3 CƶlZVUU='ayiܚO~UKĞRvx[ toxNiUߒu/p{ӿ& &jơg[PS̊-9I܀Y4)PJl4FpEP\+ %%S g'nY}fˉak.݄p}@UyܖZ4bc ψm_cQû-5SI಍+27 YWn+Aq1T1H:-*}WoR?R~i`OL#oۅ+i2 7iZ)Y-j6v~Jq}1JI]&B݂W昜3NÆ=|8VPqn>s>4/G\5m|~SH|wR3 ΊE{;h_>uPiYIt\r ۮU=T7PT5'ގ6N:}| N&<;(:l_=KJZо[Ze_҄gLWM2[#Oe%6k\ s<sUm;2WeҞ*T\KsXmo.IVfS~)2O/~wSW~^2':џsD  ծ΂1x镦:g`J#ièD(셥5q~x(z+_xoA=]1K^2J*grŤ͊씧j m,M@Be6!1dN^!3_o%&Ҹn^d4`uC %odz3ABD鉠ihR|2YIcB`v2%Hm4>RZݬ 혊6FIΑ4^̹B  -W&]rd`!|E`6I4yK54D '4drP9`U("BiA Dxd\DF(YJ:_i]cApZ&, gΐY$E 4Fh0;VvU-:Hn%y )s~{BaA|LK ɐ`MJdH d1<8_hH(2llhTG #scb#%SOpP0#+[B`]fs1eΒ vJQQFq2 ʴxGr<8\` (9HF \pӤalM.دXAa\pn, JgN e\`Ȑ9+)B s8JgPQ|J-m =bL(S|a>}Zw5-g^soG^.} W;  2@X 45VԔI峕h.l^0\A6:&,9=10A? P @TEFEA?hK^Z.c8+dJx!e[#9ʮFRH 4ԁ2@!opҠTGg PTq"EM٪,pI\g2V 0<ȕdb5܂ )oOpώ,xaŎSbc!C w? N)Z:OQN(c;E+B;(avu;\Һ 6kYch}Mb?cSNZ&Z֕D紣TQeʯSȣ>Ӕ E( ;Np( .EƤT!3 "T$_NN7Z'v̾Lr{~_οd%䒕\KVr7L.r0xq]^\[)P?O-o >5[^NO+^wezX C̢aD-hTZJR@+iRT ځe+h ""RV*JZ@Ͽs$sðRbJY)1+%frĬگRbVJ̊JY)1+^eĬRbVJJY)1+%fĬRbVJJY)1+%fĬRbJY)1+%fĬs{:>Lh?~7Kk7o}ZMV֟$TӴwGjN+lW:ZM2g\&wUðKq4gDbW&ɒؕ隳,IhX%{$vR*mW]h9)ΙԂi?٘^d1D_jy+`O%'n]<p\?Jtju'53>1U4gtf b$U0थ}vk';mvnj7A1RB Q;^c"Y[,M'i9%t u쮳l}cFyI{#4-)eАOax+-C%lrݹJrS\>hk :x@e}`Q%Q! Z EKn0h!3$+8"3j "0 *OQ0fIr_x(F(m'$ge' \[ PȦq1c!P2XiI0gE8l:Տ71E-ET/C rV2`3ge0CIYt0R1F9ǓWBTt sC}~m'&LJhJIWW| !d5c\H ;dkY'd|G@VP Tщv0*icڀ Gej[r)L3Ps5ˁW.$U(CO+ں}K=9u"O|vX:ӊ|t>EƘ*L 3h4e,Q$/E \‚mN£;x,bLCκEylˤ8۾Ґ4D_/{@T- jm+!\t^l裉f`]V7ܼ~+՟nB0R{Y^f+9^/^-z0"e:  ܧ^EMx؛ҸᏩ ~~.>ÛpKtUbمܥݥ*~7OgϢDnku?Kߪ^˜C89dS5_IXm9TXrߵćsWsO|Ω5C{W?h%UF76;Vi>jaW%GOE3œWFևhyw7:Oxѥϣz|WVq/zyȡԺO]lv>{1OY_'WLqͻ]jnJߟkm؏ɝ,!&9ik41gRlek2HZ5| 73{M [3=vXI:ez=;l1Nv=zguP*G$dUpl |1'p@s]Fs^2߼.u*{O6,l{H^WJ 7VjH"ɲpU oF xdx#5J6K dtfovKi#VvsH(^rN &}Dlb _[5(b!Vy01qGX;ZnMμj9Js^*Jp6Te;]i ^)VXl!r{FdcմXJ̍uL5>kJJFۈ|.VRM퍦)u[#ڪE3ׂwm@7Z~ذ_:oX_.zeZE`~7LM0|Re,:u}WLUS:`k搮-2b>h}:@@!j`k.PHx3jg>OLj9+%ixOq>wlLZ`zYR t3ݯKa2m!XLd{Hy P&柯n{98X `0B,*1Cy;E<|'YݛgN1 myÇnɂ5D;5Er ڿA~B[L0S!O(c :;]|2gSo:KZ2^bƖpvTL'c ~I7$!ʚYIkk(9b;Ϝ&-F`ΈWg~t9zJ0=9!Nv>ֻ03ZޜJm`yDzNQBɔ^HeSCx}Hc (WKM)0qZFe)A@`X/5^#NKWόWs%xp17i}$yI ZTi]W`r]Mln=b|Ͳ,|;{wJm0.V:CHꃡihA )ƭyHSC<* 4>GpvݏtjiCTxG$4ԩXBofEwZI%ptժ (O rQtn[~l *Xm n0"Ǒ)=wG7M8hl舢B/=WDSFҚDš8Ni%7c̣p"xbF+[@VxJZ8(B8_uPdu@KF?3'}:QM =? Ͼ5<_LH>E_Tצ灞~ $-QZpe?v~싢! ɕΰs$ySDnٕbpnlpssܛVÀP& -,IVwi3q'h(ucn6|e`>5_—4[YDlNG/8.ה3ʉZ:gS&׏_waNKc#L4y?2g0i|ƫɅy~gBMgnN/v;W㹥nn\UU Y^&z2ִ)[[Y Y<,73O,oi} f0bDtFU6:V׍j!(i"a#aWm3\!y /@o/Ʒ+׹r/u|?^yzorkQpX.GwGuGt^qEiT6M뜣Lxv)95gйf6&uDn@~U>\rٿMGBpyIM;IKuW~68:VnW^GȻAu<ӯSm)ԕA/!DDVy᷷QDPqB#|R.4Bk&0+b$+穭vQ7fG21G31*#0肍^JŌ]^1U$:2xE#z+)u",vXL0>>|hbFG-t>:~6yNf&G@ {/H~ cri)srJR%/)DnkTJjWcY z )Ah5 P 9#-o\"ͼN[] auΠFm6߀.1&aa:gIM2Vr@o (&rx Z5AR(M~TZrV.Ǐ-),8i WPKgd *d1k-0ԫќYtuSEmzhwnK 8H/t8f/nvx2t5c߫va_F6ԯ&I)xİ,oGm mޕg1 #.|xgo5޴cv dhzi Ԫ[6'q!.3e34gΩLjJ&lmw1Yicyq犡WaT#-M6.[|*T4$S*i/9v&>R3|j1ou ab*.VmaX+s|5sJ"8]k@{p8[~zu҄ۼ~)_5yz@d/z_oIw =_Y&اp&ڼ" D2WRIkV}FWk{ ЩPh3Rݹh^ʲ,,K)Rʲ,,K/eY d)Rʲ/eY`1Zȍu"C#gBU ?VKl,H9=[Pr^2>6Q*3//]aόQM`AgQd={; pF0,ALi8 yaRh m! ]#i2#$Űagr՝=,gAt=e16wGOcQyEI(n +m֡hTy MqRjp @Х, N¹8wZ8XT|搲H4H*t2PeTHc!.]* 1 JB!DthQ\1 A*!lhϲu%y GϴE,61'6r`%`F,E!I(9YB>}h;r*Pk@p:0!hZ(!lMrnxs0WQ/CKq~ƥ.\‹y<#E"eMJnar0CO6W?z ^g9{q;̐u^ÄIM#u0C."6\0f ޴ݙ9z,WB*Q0m'Hy -\6/۝gX"-O5bPǾYz-\Z3+DPc*JH6RdGk$(%Ip'^HZZ7`NVg \JZ?"k@L9Z+/f'rTeӫA8v*~pⳫOEkY||k0O^͂z5JvR%04#/E]7+j]}h7ckRߜn~͟jG+ b]&d(Vl+Zvoj*U&zȀ{(F-V!Efe FxbLdTQ.ֵu9!檳LlB}q!6bl ˜w jф૱ْ H1䨘bbq)9b@`ꁱ5W#FH{L=k=(n! 4*DNA%bbQ31≂-%ϨQ[gS-שRW>BDm BDa!v(dž_%rי?>1= >6V8}!a/}uPSty~zʸot*N߿ob`̾x}h(OǚL_ ~6BHYҕJEp1PHTPި@\&IB[׽z}N+ab朇쁽^4o^_QLlCgm.$F܌9H܀<!EJQFTrBUAHnr 58nksv=e;,P}nM-y+w:^CٵX<<\+9sg="t ^?1FP|cP^ *5,v*ݶR};;nǣtVa#/i^٠avx3ɻwjLlqM=Oh/MռTZRm=z5Buw(/ʊO.N:܍~<0JR*!D0UIu!K9ψ6:R[/gTٛ?x㡵`Ԃk6:{BK2z z gsAZҤ~秳_.Z(__MfsƇWCw&8k^u ~8=$Ci@d:R4<%=J,%Z·@m~ܕLb&޿ď3^'{0_-MD7H&,gQ 7@tOY|ޟ1tq+"QI'bͭ[mw"J$./N{ !(*khA(VH,^QcjȊԐ 7K_?rgTCukj7?@s (C@HP180pF9QdكEW_ހv}r49*dB6IOЈe g+׬l2L~7y& ;zvrvlUu]$ab~7T叧40Ԓ0?5& 5:wR4RFJ4f lwlN+f™5}?{=|ZKMpu|:3{,|2 g9yê)5U^,o&&arV>NlC=f1鵆sx*btrC/_놤}@xާi <LEB!ɃW Y_@((x#|UPsRDvOu߈cyeAP"4ԒNGH{>Ǒee׺ 3ދ{>6}/r( {kދMɑM{ Z JҋgխN!2> o3S(QWB֜}@!v͝VX|`hsCI>]1g-'Qbə\\egqřewC<PP(XAZh-6^[aijم݉/5i"禈eOf壳Ynm#~ ]ze )"ĎřC~zt|hdSgN8c]}.[+;L >(4MC#MfI>+Ƭz3Ѓx $>V["Z6+d8<%@C1~ ;Ѕx/{ <*džX‘Qo.3 ުN(3*t,(FgfwFM8 ]wg,̯R9"梱?[0 =Dтr<{Uq}أuwՄp!*lET=A,Tc0>8KNTpQOlFGt1c ^ѥh`NFf>IR <'sd<{ʏ]bY&^Ԗ=oދI]R,om^:&mLưn˰e 3(iYWƧIɀ'75pH!%`*\޲<}෭rw;r+1B"Q`gM\K^U|mr~3P \tp'+E"" T4w\c^JOwQ? (#x!X/lCZQa LERB)FWV.G]Js wc5C1x ]mv[Ύ 1/:wom[lfL=hݧkv6/gV'.B۶W2:qy%\QpQV9on~Q o.Oi{vpRFn -^6Dwϳs|C4 ǫUDuƚ C7O~b7ivtuq4脹=w/5ynU\%t<#ժX]^$kwweq E : !O6U%X=軏-DTչ@j"v1jMTX dk*ElWJ>r y$:Nn>vfW]3>e>\ -Aw&߲}`y2d T\FkBвi'qdUd.5zoPJIɃM WN_cX]R,mN&:#r#xpƃ}k7y}+ޯ=',dQJ[&H-s Ri=Hr R>T6 _19єŜ[R0EAQ{(lKoz }.]N@]WaVIDYQ5S "(ٳZmWRB$gFvC!!XuAI(kCuC)m]йxh1 _.mdhчKL$,ЪVJo%tE+U4VX%TՎ댉egݖ#f`hQ[o] k&o|L!)鲊U[WeA` iiT9q@O\rN`+j4`\`8dOWWbbA))F" q4ZtF?%r. $ulָcE =dCQ)k0 .㠺vv׷;0^&v x'O-)|Ui(kKͮQTtR?[=XSFᲈV„zz(" >?oxYLQ!,_bMuְǒDœ4vX.S~1n*8P/:KӦXS8]/ClMoA["fD [$qOGm8ͺ=a_?]qBbhj”=z$)/4UEZ W%Uu%Vb+B%TR}1D )63 z+Vwyx `ݪ zV!a#I_qsn0i:E4j_<3fQS&RI,>[;<-t%Vh; %jޅ*Sf,QȌjҭÒ:يS{n!翅RU]Q| jr*E0i&vnQF E[y=P-5(V}RLH{ƍk /B{A6AMb1O[,g(R/,Q/ eyyĠv8"VZBo,$ؐEI3{sj#qH\hf(EN[\mx޹h"ox8Y7{nNϕFu~W?g53 b U=t#{ `[>$wOF},"ȡvW8q[:S*3l?Rx d.]_␜:wH>Xm}Pr9nf`|`w9v*:)Q [mPf i :1„VCia*(B9N`.0 t#rP/%ZfC&eP,^ѭzd""/m"(ָ`:kBnǮ?tè:S%(l},=rryz) /k馥,YydcOLDS*XZYh64A7IH >-n76ӿI- -i=Dd}yޗ:aty:tY-e.̖0[<a.X- %1 \L5(jB$QWԦ]`~s MD}wF-;%.o92-,4;#HJ,i6U> ~  o B)>fW@Tga9I+Xٺ8T3He]SA&hh2o9cJD#x U\ոӬ4#Mh'U\%Cc.h~P@ySZ԰=q68i7 -$DQHՊ,l$Hp<ڇykQS*$RZ+928&)h 7- OBj󪓎_R6HJm@ ?>F؊\)ϻ,vK*8+ Zͭbw鞂( (U3Md tV9p&YN8ߔpa I ` NQh9t>*~X<od[SIsAo(Ǘo[x/ENS@YUAo8fW@vTDY.wEؘ=P/Y)C֔HA z>]s"Fxsa̙.7\[˜a sI3O3iw=VmӝE MICG@1`$dDč̈6ZvP}0:v*b݌@DR0!S&1j5*Ճ<,YkY]g*HK)C) @4"-N GU:!Zs\ôEQ4Pv]gP6a yB4T :{Pal]@K h2XbTER6EDM7{,|-[UN;1='G1\(Znq1+B;$uKc'DcIj~oyƭ 봭thԞ^s&nX17=Vs ./ n1D!%3|%gS4a=)DzyْMv]ƸzU뚵ăTiAΡL1F: lp6HF7g)Oc.*5X@Uavu1 se'6tuǮ/ Yj >\l||Cqe߆M.|wQ׽{$q8g 7ͅNnNАBK&5_T2DFUR)I8>gR% $AK$(g녶!' Di.gɐ˥W/M獜QP1sY04\`D$p{Uխ"WZu'ɶV9*F4gBԱtF!EaRi*F2Vݦ4~@ #SX O1 jlaN 9qF[]Pʧ)j2nf.]3Ok2sZcf~ۺc=\=P~_`)@C^vhnF``#!KkN(d_5q'#(8:1k6Zn ZlZysѩ^YXnm~$<jmsnA?/y$~rz lwE= <r&56JpQvomv?Q]$ RUA"JIR˅2J"@хXI{kwdp^ݰr++-ʹe;4;mok+Q;N/e$2甧RQ`+hj|VkA6$VИG"h@QAsglrD`0p*:dN8)x8.qrl髵WYm)YoW,KgSt>A \%?AKĩ'()oaD* JTK%:_ hj1FkOb].yI)[//ܹkmvF2dv'A\R{DO7o6N Zʄ2a^!X&HSE57V"FMsFV;w_IOi~Bz\qq%SX>HUO~gCe_*FɜYIsk(9b;Ϝ&8e>2Դ{],tg~NyDzNQB^ɔ^HeGh !H8hl舢-ۂ'״j#7E RJs&HHPd1hff.## 6~&au9Ni%7c̣p"xbF[py_Qo%5.y5 s%MMl7c |kxYR.T \˴U_?fMޅgQ9g@!ZD?N׏@Oق!|_t.C 9SD.ō@Z]0p "tu^ޫ>@0! H ˥CRb0*&gط\e{7(4>3CO.ᄏT%H'q5Wu.5f #fZYOo`JGIYD݈SI6T>.Sǒ#o0?B#%wЎj{ k>*,[Ƣc\lT10 .fdrH8+ʦꋊ'H` ZGK!zƌƁZ1b"iZ+Ik;11VIs(yԂ`C ;6rϖ~,S͔ |:_H rkCsk@q,廆LJϢJC9IyRN #HpCaL#!K%g4%#B"ZQgZj*B${gs*0HrNBR.z|.pI]yHӧn5 U= ,r1W[j\mrUX;ҜbqD! `&5${<3-'(s댗vdC!Csqh.š84~!C3 i 9yCCsqh.š84З)սZf⽿wxI(yuT+wիyߵ4.6&יR9ɇ(ĝf6jooQS th=%4/Vh^k)waKK^,Va, 4c^=FZ\{=6\{kosm}?&;-b7Q.ԇ€Ptky|qؿ.|\iJ\;NCҟ=Ü~rVOxypk' w˒I)O\a>)k=`zjKxy㕌Y Jۯ BӃft ](=W}L1ӨXRP~z?n/ ǷݹrTϡ>9-uШdޡ"lLbz΅Uoӛ6kUEoFg|3c̦fۘYLol@~\cj;7K BD#rq'p:pCw'(o9L#gY1b^Iǝ3Rm5OV ~9"WŰ?jnM0 ;U`v9B".o?L.EKQR.EKQR.Eߚ3Y7]zli%HaR0KwP1͝^\}yV*85nS3v;^B оcD2Y+냉KM-a$EV @H?8eVSw~X:\ii}'tWGWiTOm)߷F^+k)L@yM>jb w MX5ƨj}t\998\jliٹze m6adhDDjifMJLY`;z 9ub%sߍ,be5uJ@ RO߇o~>#wr1efAf@ # GXr}zO\o1ׇ"-(n= 1ѱó?ˆp) )# dVKX3ɚp.Xvnȣ۩?'H]us~7GXAf[+vJۉ*ꅶ9F Ib͌dJ /ВZ3cc<o5m (WKM9O1jvea8z2|y/V %'< TjvP \>‘r4PC6utݛ*5Z.Wב -S3TDs/fEwZI%pjU'LlN;ʏ5A%q/`6CYgu+{q 8Ec&FSAƽf:|`>:h&usڟhچ0ՂGn*# ]]M~+a@UiC(XK ń݁ooq\}1QP-.17)tLs-"k{5kvz07QNlԭR--'4qʰ\Mms :8Qm{Es{J?kW?+?^ߌ\*f! 5j`N/v;W n팓{?ߌ\ )r dꊝpZS*M_mmf gqYbyLC_'`b:ݞhÑ&2tW:yUzV8%U)U(Lx$ḿ|M~,8ç7D~oQњEob^sp@}|{Ƿ?~#&]||k8qi+AOO'po.~~֢aTmm>[KIb3Q[7=rkC=S>||{ѿouWJ;5sH]p&~zZ!KPc*$#wkȇWTxW9^Om]h+T1"h־XkkKG0$BDiEg~{E;'U:OʅFzD&@\J"I<Imȼf0ϐ} AA.T0HCZE#G\:"*R')BeUԩfD]YPGCk41h@[L{F=9#ɟ92jaS]*W#RMn95W!l~Iw1G`\*o2LtθpUbPGupJ+S2PbK@]D4 m$.J:~ew3D[ Ad;zNv W6-3]|By`Vr4*U F1%lje8k5f ,2d5ͭRV{d8éc?)"(R2BVAD HSin1b XaZ$)IT,)vGE`.EXS{+%Zɞ̢rXLG0&5P # e£=H( u;c\&Gu eچ_MH{bKKVYqSFkLr̬!#CAP9F JlQ(֢45N(#DOp).}Ԗx93Tے5p6KZ^dakXY[*)DF7ɝ|b=Jks|ܨ[F~sCiN5zF`O@s$ѨN0+D{CJGԵ.Dj&[pK҉ȅNq530  D9ZBHXG, i/3e{A<FHHubKFbb$S, A9T aK3i2ӎv4ChkUizLS-"r6~ݵuM\DL vfi/ABM-R~/wR@+Wo亮ox03,M曃,YydcOLDX,-,xGI4 GZg{fE#=]m [ZZpe3 >fCk '3wo^nU֜gwO:!eb;i<ɖ\]{Czt1oǼSFՐjR5k6 %L;J3u $k_վƘXs@yQZT#cޖ'-F<D A3RkD  R+9h=K"U83gpKLpCjim<цlRi{pdhIR\aXn_-`KnӢ$JђdK8쀛"!,oN;D:å76bCNDG4( 6h$()fB".sƁ.ݓF!r—YnSIJ 1UaD ƬQP'aR: FbPPB=/0+!ٻ6$WxgwU2b0[ % J:,SIJl~n]oWSY{+ejv u}]\z财v+4|:`2dq}4*[`\#MEׯ8 aQӣ3SrF Drf]mW߷ۣS 5#ܺ7NQ] ͍x.Ee SįsKchFQoTb4=㦟OgG zCPt`5JTS/cϟx` EbP5L(ȑh’pDp XFL("'xa=+6zvz@~-7jfNH]$-mԭ{=36-y޵<1_n2ȍCˣknmB|0 )cmy^ꂽ~\l]nZ|q~W;/]YƣцͿh&=d#~{; ӟnqhҬ79Ҿ}[}pW,l}Tmg!ˑ[͏+JgDSbΆ*P%h9:U"RJxT E?N>CLe9DBXJ=.  4q`%b͠OFEP։PWWu1`]W:q_DrZ  ö7a_VU sU?peQ1ۺbCl ؼ/r+b~1pk2?+w*m\BEm.lߍB#A*jB *B *B} iyFt-0lZyR ]+V|e_]d徵ӳŸ'88a<&)eZ6k,w׫ޗ0x:$5c@ĉ h \v_kp-۫v_ 1uL`˂) 7K*:_Ѯ_5xˏf.P0ca:kÊp1iPx D;pnpЮ!Ev&$p..ap5nՊJݒq~ǟ6ZA2ltk=dnH\%׹i$>#|zM2#76FZmmQ/3nb3eL1س)) #-m"@㤃&1ɜhmOtB$/$4ovɤAkҺ80G]^QB"d?ݵ9nx^ F=A_]nW3\^)…GsBA* xtxv37r <в SW,AHkьqT42FOfG+%Ҍ 3dAתunmZVU!kUZVU!kUZn"?6T.BۻBq#n]h2hi0cx#,q/?*'F{[E{czc3+2y+u.@3R3`@u:)IUTi"1p֋\ק(RR1l $ItIPDR=#s0ɷbV)a- kuNiě JmJ"݁ӝr{; }}@ibۏDHA0\!5hy/,wf+܌3lǐ#ֲ1)՝M&V f܎Gy:KАH9w/i dzm gi1xWA2B[Ҵ\ԧlio.* ϻrjoAVm=)+K7~9$J=6?X(7h4Ba|7+f CZ\Ղ222_Jkt8mi'ӑzJ&9=?=}nQ" R ā|O;߾g]{:[ZP=#7*@ȃtvHo镔̝gjSηX)C11`+RF QSR$BQ1S ؞:%4"~Ĭ9\m6:E~{ml5{mVZ1hI@J5gS s)MG`DKQj^K iN.[>AK[)(q$6R 1끑Z~8Ib aUN:w1ss {l=KS, Z恅>4flˆٓ&ܧ>[x_~(@{Q##E`\&T Os̓m<_Y`KdE+ex?r({r}U*Z+kÆ 9 SgQ!s;nߗ-UBPm$L`J9&=e Dp\ zx%x%9sq>q :|T75)О! *i:ޣ9Fq67JAOkS>bSM_=BrYMgǗJv2OXDqFېo9d6io~4&#>At,8 m+Y&!v=SXVGFr}0 8dYJ CEV UPZM mH,+|,63XA@IG,y,9_,mKv쨻Ef=."psIa")B H -o<.pEt6)K]LLCI^$\3pvt%?PrM aOլMwp|[ [wݶmx-WF ,|WCRȼe>»(HJ)P뢄WAUWU| BCsMYU[/$VBhHX8)A.+X4wGpqxWggѨ'+ĺH RD'~[3IJց:y tGbL=?hڅ@yJARJ4LBF!1G!xS֑dHDŽ)k;^uQO0HAtygdA2Pѥy%Lj}[~Nz>sjT\v<{rDQ=Jxko8Z}k5 r[tTy;8 {}j RF (ͽ@J5Ckdž@ oNśck i>737yŽ_gDXUG$D! o$˷ /MoC#?V y=E0V{IH9 PsOxfof'GQiwQ[x7f )B9ˉO,>BK5-فʚOغ9K{٧ŅwM`[Pm9 ߆2{$'_ga;FfgB}L'YGӬ#{&ڔǓ?M5Gˍ^9<ζ,7rRluMnZ+;{Iz4R'֌>Q|k6 &ucP bˠq`~~?tﻏ%Οxf,A៏@Gpc꿾SS+L\bwW.aޛj֘Jԇ~SwӚITKyY$d2y_aOBCPTZXD:0rC=f=*vM>̆[#)틭Um~-p] _$)ڣwR$,7b PU¿6^rK%<]ՠw~.rStФlԌ1W&fhP*}cgjpY.W}RLAC&VDȔblDŸdm)ȬI+! fZ;Փ2sB79/|(emOF!g,k)!ر8燬fWm-(V*?sӺ?{6 gvtW~.v“iG;!@3ڥcbWi_Eu'SR*%^A=nըy?C&h(\5hZ/(lŤVp݋Ieh=V(~ô_`{9/'5R|?llrlE훀̳%[OKmEHlL$P^9LEА3,|= m2%j˰Je^Ϡ\孍#c0W ]GAtC): c2,ycݵ`GTalEUx=CH"&G]b$@F) 6&+,șr;zɾ`ݦwLg$\:)!,٬\){=*u.HWDbS: e `DFil=o^lɢ W\3pv v ۂ|PuI/!.XuZ6* ``ijrQ3(Փ@$Q5-J+&YJ|%Lr!hbQ@UT-z;Ej AT~$tN񑤍ָ"<rRQ! YG'rL:ÉWcYX׀wHz>hWR;#yE ]-"3JX_bS[wf'_P3y(oYi"SxE1BW7PTb [,I*P}QyP`7Yy&Bac:u\-N8Aϧg:SGq8} ~iiVrˀ>6nOqxeO<փqf3mgtA$JFEYA;DE,Y}/;e]̀3&\k1k[YE>m9֯^\[5ĸg lCRp{DsR֜SDC3"R@-J ^3ܺ{pEɇmqRmq+Xrp&R:꘢P1 lfkH3 pm< e- l‰)_C_AP-h:Eku`68;XgmtgU;Óܪcwggmi \-Z}MJ6~7>.QnRvʕwhmPcarv &тjIF9# =M7ΖJB@2i2\)lN Ca&5g 9v-e/3x(/;\  kW_冰͚]N/(rhh4:NgcOFdTvAU g-j4.9!k:y8r oCBEdeX2 &+}=|57rd]3pq2NkpΰPvv6xe \R2HYY:2ڢCNeZ ]n8_KƮҘw ȆP֨cA%TB)K$ ˣ Ѫ`IF.hlMOi=) bUOdC"tOZS(5I4?l#vY郕$Vkqր^XR|\%EfQ%z۵Z;-#q [ 5?{d641 ΋x@ʷk(J '/U+PBe`SK)_V˗U(˗}$ɗq, ۗ]4(:8MV^5}hB ؈.>; $u\.{@f49>򢥭a%t>-Z$_JQ%cP9BiQEW(M (sp*Kp!; !^Wc>a[<0lLEhoesGڭї76@Dfvv:*X7p}-Rc@ bL^?iLX"ftuKZɀA:H]e=-r'oĹzɬ5jf٠VcC^bЏbQE—)ɸgGqУctBȄ$DeT`d QiػG[kK擰+j*F) 0TsM- :Om]Y>NggG=Aq 3"i5{Ѡѝ19@<>3i2CQR՞15t {J ]Y y隇POjq"d 4y))00L`"'J7i졇Mӥpm-خ~ʔjH˄Kn[s:zfcQQlw^UOZ7ǃ@Xk"ŀG7FlPYI}JN.{@T50f{blPKeT8OQA56@aeJ!!:8D) VY n(()g¡2:;g(imu: b{(}#7EЧ;`;fEq$bq{_Yk)#ى7y,i^,Q0vewS=5l>ŧȇU7XKŅC o|Ha Ϧf% ʡGM,j W6$Mp4 -A" M1}Ki,i=p)54QIb$clqhah cj*c90'm 2!e0G1.W1#3P.ɳ.lA.jJI+dj%)&cxyec V-+C,/~G>mAr ASMfjYkGj]T0:=#w+ ]5sQ=wլ4vw%VVz l"}s"UN?}S:6hmł& / VT)X>=Xx{juL9i@ ƠPBQũ)'&|'qsuoC@-Ժ@ˇJCܶѼPjZop&/c_Gяnv!Z;r{ ␙},9S 5ْ1d%tz(uֆJa糆:R5jZtQU'JmeP[xHy.&,?T+r'A+-u 5OZ/:-%vOn\?Lg{c|vx;Op)VM"A(%Gs Ue!NL5r0lt2{n:v,5oM0"?D4WR& ]5>bNZA6B_P*)|*N@QC*[h T/6"'aJ^4nhe6OB"Sh,>6}sHr|8l+YҢ׷dxW%[ϥPW_HR -V3QF' 3`d  2StA]=P7KUnr5 /N}\%L\2ՊKv XdSo͜$;YQ|aq__:B|ᣧ^'T.[-nS\}JW;`󋿟=A'JUZL&;jb(U8s0]fԅ\R\U%r7_?{#*m|c-[l\ˑ`#`6|ل 7Sm ¥\uUG{g ?3\;9sy#Wo O _1&ciː>t1tQ4gqҰ7in?Bn Wn,O..ӡl4l8 I%-dys"GʎQMXN[;,ZN'E>]}̯<7] ٧%3 ٤nAVa&*j1Lh!l9AxG|K6&fDv2eϭzx׍ŢokN9st$!(0X"m u J9rd HS]=99 Rh\ˑnc㾇tHMU߈zGjoYK$}dqTfQMfi7+yCe2P@`SH@% Q9Č,%vժdgj5Ҽ{r+"]^ؼ;bm{NVmSƿ}?=tb %)WebGb3٨:7laQ$_6 %y;|LraG 5=Im)S~cjK`rEOA>椡*2^k`ba,d#M[AX9kM]=^O:tŝƔ ŠTt(w.ng[|*/C+rV-҆(<{j7w if[dRprL6X5$rr!wɣ!XO(*F:R>bm.HANjVQb UWl62 r,^:h*e\JZBEJ\,Fq0q͜q6o/f+O֌*c5 a*2c&2?}/&*}58p|eLMdw *F'Mq㖔AT-Yh!IS=hdH2(UqtquAe@AZҁBB`KU9-FNeO bSƔR?(jNh""}iEDV!r_u.b+\o3ǒR-+SIʓwdS",p[&+vG?xبiv!D\xC׸MϷY&üZnYD`%&:ȡD)NVIlr,IN,Ϯu_Mq5JʧfR8N{З&oO+䫟?V3׋ܣgWj=\OOؿ?采YZҩ>ʠT\-sʈeSNbIE8_V\qBrR!zYXxXx]]6ZPbΔu| ˽58|}^ߟ4s~qx;?9>`VwOw?>\-'O'SgܾIw"moNo~w"̬kS/g/N,?NuWoLm܍'o\鸞C 9\L~2#&F2IAոtRc=.HA&ݖFe4 6ΰKR4R^GT^kMŢJGƛT kضk%%rMTͥ) I3*Bgu&%kVw3zdʖ,}ޝrXE]*}!+MJcCXYԁilHVlshZDb,$P2EKz\ hA/o Mcdbr6xZcj*c%^628灩Br2!+]TPjt֥dCTC-AC'9)1B°r~1[Xʥzr%( n!%I5蒹 c{4.+>HXxL=2<{rRT1HΘdLVVRa؈06+%I+{X{Sw_bj[;݋#^xGݗ|)o:|a1T@H9I!̓]8} kQ+޵ovhXv6 EgKDJb sHq2{ti]^o=ƠFĚc*UFscw{"/}l&Woo"ݝ^[C! Nɺi 8UI7FΐL Xœ&vJ86d1&OWU;lpd?)io_?7sw`wmy 9498%a Z[Xq9Pa'ؠ 6`aBpBs5ZΠBAU1ZѨdi3BRg&E-EC)9"X `j4v3?f΁ "ˋm4hY Z5Z-| m}ݧ}kh4RSu)W(!;#VScŨk*͆7}]&;jb(d jWAua2(+1W:90vEN_ܻK Ͽ cw4>?^[9Tjf~tقI")Q61[U-(ju2Uʥсu)6i}&5RJ2AlbPcX)iAPդzf~ɳ^[^ P-qWxwL%SdyzF99 ͟KNf璓YsrLҠ0'>?/dF8KWk ~zWayjƶop]a!{ ٻrcW\R\-J"S0E49H\.yT"kYT&eU}o쐋ކJ h\ϰמ /淳VD&RUyKQKñCGBG`'ďn$-J~x:~{ǭV$E/ eil gJ!.Vo":HiWGwV|{Z_c0=1})X#6ߌ 1?qɷ C}JV{s ,@Wg0b=b! sk{>Wɇ"DE!gmhZrmK>BrJk!>*9u@1:JI|az<~+@C>>N#8]`jE؄cNA#HE hH N{ rdAdEj"`Ы"N)L>tLq$<&Ht)BCQraj:OgUIpƽi{3i;3L~U1sX-Z;;BSr?Tq9VO]]547G͞5=/76N#=@uX{!YℳKH.V XcaenW8IqyjΚUʢS1P'hfK;1( 6X=r_GALд(GF#FJ2{7k0IKjRIe]X2ڌU."xH o~zbB>/{r8Ŷu\|s9[>ۍsO$Ɏzp#󘵵|]$AڤeXY!U[p$c ښVGZH]rE&tvInBTIm9 )yR`i@wt6)K,'h.ٚ8;R=y~DFImD;#1O~Lsu]G۫n^_y-K5tΫ%Nfe 1K9'x]H+!.6QQy"Aj|a,>ꅨ{lcQ&EiIk*0x(AGtVK+D . `{p]̈́{yv@ ׃AXNߔ&57tDo+EeLGaƒNĔ!(A0v]Pht>W&U Z2{}8QLY(Kw]e^W {GQK{DQ$x1IW/[MқYe*Q?oz&x67@oR*.F+xYo*[t^y.0ћ{S.AJqjJ#zc&,wOTXp&)GRp;U&-iZ7)^-"߆O̓>/]f;;ˠq9[wc@ß?c[~_hO]ӒP%Ы[Z׌\ĹdmfXOlGwm \۪[ur]J睽Ij ͳ5R4}~}kGw]zu8gw=1gZ7TjhN6?<%M?vǷ~?^H_߾8zw"YL|(G4mjڛ54ЮmvyMy2ۏ#Y@J(^_^_ԛ䍡NM҈pzzi.ⴊFߓJs+&@Jjwᩩ'c]I_lԏ}H@V4E#hhmN, |gA,|`YU471s#\rRAZĈ\xȉ;lu b [J<;Y:23uF:.T_zKA*X+!Jm6B!%1xhVio[eL`KhD'mTn:˨nr-ItFڴI9d]RxyЎgwv<eͧ(}j>BGQ ˨wK"+ mFyZT%r߶}-q8qYΟoZh4 zEYn,4Kmڃźz=N\}^֚zȏ+|D\/zZ1l!xV}rc\km@60LVB%(Z ,:'({Yd[Yuc|e"rN!$3^I'INbq&¦,=]gf}m:DdrZ.K7x PgZz)+?enZRi"0ǍcbAIόҿ2nhAGd"|q۳{־, '5vRa(6r'PǓ`+ U]X(Ĕ“HW(4K_8 4,mkM;퉷Gvp[jwܯ? 1ߜ bX꥙MLZZA)#(Z^rjF^ZGiKu0*@MSG❢Wbqn<_=l1|wAKV=g .i:+mHѴ}%ߜb_OZ`5߰pDRi0ody,= HJH>PIpPC 2mT&]s6H)HMۄ"jy_GE:i06H2>lM}#)^~70kn ,􇾹itu3I <| VUa^jpznzz ^͸FmN,FFCn'_kĠT*ҌJ7٤i8-lIGJ2*ɦuj7  o|߲C %7χ޸nw}eΓFۅ;'WT<λ'^o7ϊzpnZ!\/삻g[i6PD5)QDDmF4s\P1VVZJix:b SmH!_IV`>3b3,&$Wl=,?dv˖NjEvU%9**F[[ JQ|s*&W#\ҭ o&knqLx?mOӽxBt{Q|"3|/FmwPzW|ZPj{_w `qGUNFyhmO/5]jP }6(odt:&K, M7Ei Uۋ \Gw t]lH.hJIO2t)Rc i4x,{\Ǹgcͥl.Tps}ܫb 0xsi|[m[v>qh %+dd}!/ƨ6\]q|Ho nDj'gYdۚE#\@t!*75}& (T'C7869(э nɯ4պԞ?$'H`Jׅ^|&SZ6$w[`nXj<"˞Tbl"h 14>D J)zqsC,/W"2) t٬qSb7քvPX ŶY :åY3D++a/NnuɻM䐅1WQiJxpuV sbeLo$;[AWݾMUgRav(~loUpە-V$TTo}[sGz>ٿbc-dS)16-41n)!yue 6jQbS6ِbgcp;sZa|_$GeV%8cU)UE&kE%]TNw&h2, !D cpEMJ{ZIL˒Hر_3p4,|?)$THںjU݄gUy[աݖ<:_Z.w8wu:6A OYeT-r>e1Vu!],KMjB:Ϟ/`)SZZ,-I XGMY)-S]6+X4wGlpq8Oݶ@h%b] MIIH çN!fRW,g/g,2BAς1TG:$]tT`y0 ALj]LYGR?G:&UCUA.$瓶: J$%ΐ D@2楄 M 7jZbժQqur(|:<^3;FՕ >odIbr]e>j-گӼˇ0N&?^/@ )k'l8J3\Q$~RX'+Վ.;{0i$^f66 ( c›- JV*s޷!}ռX1(O+R{QH9 ֩9CD;y-zzxRA+g?Mhx/4FNg|N|#|4{0zDcktvم˂u`j0|hxi/vH4=;mo;I 3FfqN,hSwb=ŏƣFO.OoYoY}"Wͺj| LT_hHX3&ypTm=98:K3Z)nyaÆǟ~z_~x?7?|}|?o?~ 88TV` ¿R?SSu]zk %:Uf8r[}ٷ7n8[qMl?~qjF<We4H0L懋l4{D ?@@e!N^1Y-Bxj&Vd()_ܪ=&/BIR;%b1YE?A/DLM6X,W0yMK>ma 5ӷtDKa8 jTa=k%!00I;u:qL#we} %CX 0ykA,-;.,^ Wn+_hsSi :#1jB jt:[O {L)DE;ʺOW)W|DŠ2ɘ-+Qr1.xY[ *k TlղS9ic C\_@5:5~m 8iٞ{6O^A8}=*y83IGFJw-Yb55~9::?y_RVEG_KAcۗe~~,4<4}hB ؈ec"Fjk!e]Rb3x%-mTy4b g' 5k^dv~ë@Р@DwME@V `ᕰYВx*!h5fxJ&tw|cy VuEr Σ!M0jO'&OZx q[\,PnXUpؕܰ m P.Ksþ0V3uOk?Rhio;XCdA j~c19hʽe3o,nuۃ9m4A;Ԫҽ i@$D6B,QK/3嵤1Hu%(X"ǺLY$5y.{SdTS1aҨ Jdtl` 5 LSEv%+-b1)/qitD.d/R=yȯ'K_8V. 05$BLozkV֊ 3ѧJj4bQ& I2%GɧnDr(>{nۑQfAS6j,Y$Cnm LD2,ye%GL0G I>`m&&G]b$ S\r=()3m! Vm錄i#Nt %N_.(WNzs$Y%U9K6*ckekCv(\q9[gLP}YgIg5| _D|JbkY)$FiMU3U1'.IK^O?73# L#J$V9{#h܁(dED-:zR']I9R $sb] &mƥXT '*%)D!+{ تĹp$έ%45R$ y+Q"YG5,ȨemK$ K,Tzj"°Ք1k֭AYdc (sBjPj!)kXcI#B"-CqDdݓGw::> CHJtKJ|Fгɴx~o~Ҵ4[^B2z+HԈy@BlL&+jU^nc#QɧC7Աmw}`beM(~|Efwc|pz(={Rlo6VZZ><_TsRVj)=DB^"fH"k׃Pbzc:kz?[ Jf~ Jޞe]?K`̿Rh9`#FV- _+RH:²M)ֲrP$Vˠ9&Wy|叮YC>8i >ONηJ OaC3MWhShVx+3N+u^+?G9jwt\y,5&g`B-$Ơb4@Ze[߹qQ8pTz2մ1*\)lNFRj Tל3p挷K) ;̎mz^Q^Z&W LkSrtr>YU7zCFӧ/>1md ;J޳x 5X`R4ZS  6T/$QT*c`4VE^EuscǣqZR.kwµ}ϵ{ OUu()W"HD/XceF@PC Xn#J:ThcYrZ"k,Jd- [5) 2*+Gȵ1R۵߼3p懭R_ہ kBx-3x GĎ%sė;"Y#H_,RY@LX~0ٰZV"H؅~mM=X&\q>9͇-Z+Mb]u+wĊMݸmD} 3 PQ`3E+RJQ 2^B / 6ByНǽYt#]pWͥDYM~ 4jFGhi$j5|;oYE'`:`R b/Rm*T1PYsbʡ)RpP K!B(wz}~?G00Qnt~\ǟ[]nUQ=ouWla'g FYe=tt#-0\DI )/}fRetPL%Q(5DJ@HuA7 *wV?{Om-uZʃĉoU;LMDEiJU4Q(Plz\n5g_˸LH\]ǭZKZtb ]mf lˆ4e2{ xbYΏLX=Q2Q<8z (!v(} gj0< X@V5x\M;Mp۞u˭{Nng5wƑTX4+ξoJ̄ r{ D3qtOƓ't7/3[/3g7YGj1ϻ_ɻ0{K.赈}җL\:ă9Α0H{eHV @4EQav&&-& ϶]7)Vt.:gQrL:.F! |Zt%n" DUMrAչ^si|uA*w]JV{I/􋽽 iƌF7RMjm5¶״iY@LӢOn`ؤ#:tӨX_VP^!)V9I* Ytx,(UXQg _m~j^C摧d qM` s*]j=km5Rj|,klYqS+Q0$"K kʰj<[B5oʍ[ںYsr(=`s,Շ+}Jg O Kp*g ~@X~ Zj(YTFEMl&q 4YҶP\o}Wo]ut ^q\B1Mkc&ʨdb"ΙT)`h ̱q߁C|s_2Dy?~=+j08gǶAq,StU8/̱!5q =Hq3Q{.Bʩw㴻_U 4 $4s%yƇ( j%Slf&'1% ''biI̡pgxIS+ =t/g5ht-TVU#V&N(Qd\O%QuJo1AF-pE hĬƍkR ȆFѐ(.2pz@ ߯y\Z1)6{p.}>ZS:"W;?r ev;ߺ" A*'&TaD` eJRJ}^k9ɉs1ޝC)8GN& qfjԀA˜^4Y$ۨ`eԍf*i Al1`svk/hՖqlUONR?lsdC[L9: F-wƊ ^UΚh+CD4:qIfJ턶,Je P Rdb,@ XTvZyJ.1 $C2 b_~z%dGw^mx-H.976I 3$$cD/̑.d 0I% :K2~|VTZ| GwRPSs,MDaiIk̨VqUv4[5d.`y:3=["ĽH4uoovu'79Fѣf"*P` DV. I{C99gФhhW֍Bݛ7YqQWcwR1*wF)yN/jrѥBt1YW/2Gz`*wA1Θl3(N@C&%θ#C8K&ήم5DZd>4fvm"{d,-2zp9 p=?~ǂAi|[|X߯|Lq鄾^O~p[&Ѿ?=Bޝ~>_ZI-WNwђ5s1-v~#9qI#||Pބ4{0U镟 ?^>x} p0q /.g+lDoq43 +q6>oZFl\fqVYF_hc_NbW'k?olNLә֮*Uܴꦽ4)4& GM* 8g|p5>{oj9PC8Æ/ Dӏ/?|~#}~ ':ۄm# EYo~ҦҴ|i.KC]|u ~K>V?֘B$w0Nuwpn#?% 5"qD柧mgh TY9,ĉ%3YAʤ|?HsP l_lk 7n$ aN2}!Lv/tR`,DP&}v3%..uo$a|!Â#t|VOdL!Gk , +t ,Uh%c*eT tjFD5sFwugD=^|"ie%i^/'l0)xFuF\c.:)K + tM0J+SWd{j_E"z2, ,bD% ]yʏ `8jltQACr:#y#ľPzmrw Է@Xq҇l(xf,h&A:DXVKyth+CQU7KJxMl+˔T[` 4TVT8+9q5v.I-p( R&VHeK)/&rQnI,lae:Cs$ u N2f&鼎 n9˧ǁPtY ) 0NS6cKւᒵAY;\v)YgLHy8;YR *]JQ27+%"dm]BʓȭiK>JHAHٓ@MQ5RKÍD"+'dr<1! I;bRPZ/|q=!Kيg'F,a>I :\PD{N!{rYn c289) w\N/}ыC}%&>ӫ8 hʏ}K{s?܋o!eeVWq8TjV{_G 3>>_?`b[vg|Zt<TX+s9Rsy˚49HaP!F)"ւ٨EIp%G)L{&! xv\5& ?3S]! [<8)[Npa%z`>k85<w g$t@>YPPL փe!2cJƔĸPG#]U٢&ԩ2xMǎ69RUGY"HGXJ$Q=̼ܒKqN8ㅐIĥ }Ԗhi]9[f? MRGRS"c"#bFR2WR s`=͚dQ /_P|,TBSwu1]pTDI2tܣ-- 7:;-OevʍtXT>[: pX2f߹@1T1GE:h(!t9QmMS(CE~o[v\pV pjik>O#wg^>Z`?}F`P<h}! }-PyMȏٵpzaD0JvXОh }Jr>c|T:ya*G4&0L/qLYds4Rk!]v;+ޱgr\2ᗤksQ#=U+ty_B$i_xe2p+G%CtQ30Qh圤 @%m M1eo: źǙm7Y,cZ\L28llJj.*Z@5r6KJxzqqbV|@ʳ}y)acR=Ŧ3w2%> ŭ,.O8'}Z+@imԒI4%` w2YN6rYI[SsUzr9/I4bCƱS>Eip\7cn|sهv~h{|rq3fPh |lEo\N7]]Ze%p6ߨ0Œv@j+S ՞ Gt+!l+-j=B5l-rJ;ِO+_% _BLb2 bk-ugۋgV,*Eg%(4%!\XuX NLn*Q'NK2[87XbqeV9 W #3~C(@(vj0P>% \ҌIGQEt o7rϦ< ׼keư={/0߶p~1*w'\KD/;9B?FwILjϣu!݋pQYTj>4"30F. #=ͥϟesFv תR$RтA\7&X\Am[ m93; `E^ȨM`t":QE$4E8HG`SSr))h3R)nDDe@_+eNh-WA`RfuB/o٨bص{|=`nxtotwD%mPG.]G$\:܋e4X{zӃƵ]~}5 /4r=?-\ -x==ߝ?}VOwwf.-k.V7ﮟ^xղr_#+w$%ϭ$`݅Wg.AMvMJ~XFɊGбȍ&ft1I'f8{Ž\DHHKN:y$]ۉ 'Nu|Of(Szyh$$ LPQ~hv~vyE2B/W}ͯ(\Q/ 31dZ⥷$s}u2/Enwͽwp[ECE|m|>fn2B/ '6 ֙ }vJvP;[۝^!'MZcE;k m;۝1m.> \QZ3,=f$WѣIk. ~iͅJ)[Z;LkV,{ BrB-CWWJiz NOYvQFV;X``ōі982jFK;H+^h41Rzm"cAe:(RJ$ AoEK`idaA02\ wesޘT]bP|n\ns)~9M/钾@UOqS$Wrq,5TC')KfkZk֚fkZé֚fkZQ5[kZlMZk֚KԚ#kZ=V#kZ=V#kZ=VxQ;HwhuZ]VjuZ]VjuZ]VjuZ]o+PJWDr݈Ha$'=FIOAlwox^lJVUYk3!T1dIРsHy-U}w܋kR\Q[\eƭxrF$f"U2}jlO rz9Arzvxd ͮZ0>D^,Dp;!O!VII茍h ){[XSXne2{U2u%$+3f&pcm J8}79%x%zr\Eip\7c9IhtӸȝ>C[?ߛ'瓋/B?W]g+fWe5pơ,͡o,F>h"3.ތ;vPKlQA=A*k-7;TQ/i< 'W5Јr)n'W Ȥ KB)ģ&[Bzb1"之&ee,IO2HAy%: UY;8{&gs8Bz#M9]"1b OÊcYcc>ƛ/n 5m4 )ʬ>v dSˏyVC/[݊"R~I/wh_H:yyy o gAEЗms rI}p،1KOitXqcB,w/O`zJr+YXypRG3$sy2:8+I"4:Q5V.d\邰.EKTd4 y2^?*kFXC-tHg$ډr7Fgާy?Sx|lgOG;>hf0uEvsHHKN:dN`c$1KoIdzSnju:S:0K &!CD*,K1L$d Xʔ-rla:a"L 6gvWNcmc9{ O(~ٖZ[X$P?'O݆c\$"\Nj8*MAE1N"dÌ)jZC֐N`SGAt(%KE<!m^> ܚ))<Z+]~s|kmHdcd` ؗ5\SBRJOpHQ2IQTZcriNT}]U]N<)5Bw3+oؤP8wcU´B0f9(xn 'uso/W,v_z/pؽ4+ZZlԆ6Hݯ{,N+hT6ʃR uTx0͌Nlr11epn_W#ɑs9hRƛ΄9 Q ;$L.cI2$L%] LT$=.m:oq|zjxE4 byNY&*Ax>po>`}}{uFs9 qEylHt.ZAgUAdT4%VqdJo*RL"@(z.ILh*#PJ 3jc"0  d1rt,E`J9͔}[k}<+孫L\ޮK)6tdʗurktkߙ7-S)c)#ȼ4 ѢU`BdEI a,{,U|'9,oSS*%X\i=I. @D NGD X3283m<NX V/S`HEDu[hG^3Tl/Z{ďUNͬݮq?+\_AQ/gBI7Qp2|3Yz>J|泫5٠+#J틍sy>b#RFSxK2YN iqpM6RT Yg3>ayU->ia! -έQH>Z4H4'F'jZ$A|b&e*q1!e;)r3g(ǧ:=GKyԃv̖{N/WU]a-t_c*fBvPaK1cy5)9BG-rW'"QEuzu Pt87J;._ m &Hr)4e39`TaY#TbbW`973 InOe|x2nczxV˘ jP58c=GPg.ÔNH*S O@-894DHG4/DK=|4ciA:R!1oWE4EgQEeۡ7-fb1>LdĝvC">x} '#ON\TK-k);NG9kيǒI[G\+ @S`Cb;K] &ddVƬwJ(2JR9JCb@z^ q4tJܪ$Hf A42#qdܬ qƮXh cpX(?:OW4X2}lr;MՋ?P`?~MDYʲyX n '\:-c(JAla\-=e #GͰɤ;Ƀ&&ld1tZzyh˶v1h j; v QOur 1J2aB$X<Y}0J+JaRpQ3VgQ!քHj/c7u!@<(0#q<4-X?vDZ%;Dܨr)QP˩B%E ju&Qk2$R֦7mg`T4eqyTRdDԴ+fpFb<jMb\+.¸.vѻ +x5֠3 C=b!0CsQw\lO)h p(^4^{V)Z jKe'!I3-9VKY'k-;폛@dV^"3Z/+8auQ/[ 8%q JbɇKM_DC QM򏹀M*E?|_G9gA߭Л.qSC>syk>ۿW=F?B@f :䦕r'&}\DJ0A)FRXR@9<hEo C﬉«9~x3A=خF(nM,Z8bm0Vjxy7@ze{ֿ;\!`vy8|`YwFN,XP|fXD75,Ub[cD=iK-6O n[L7>^N;k=pwHJnoo#Hʐ4 ͿB+Bu+﮲7,ÞIh# 5ÆGytWM#bfIj%1lpW{ԫ覟[5__WxfP}VFz+^U )HI2ѐ VV2ZpR[K ORPmIA;؞rgveMv^(^k%*c2c1 ܙfX`NO:A@xAG#8DsiFPdt|)r qEmvGh^Aݧ6tZ}ۉD)Re7PR.x '3L(E$-#kRx%(*Pe4*@Us>H+$)[3U4rU8qTK,FM9Ce4˻vۢfƼMI]pq(Gu@1 1ԍ5C#'Q$n?/g3O#wCuf("Vh@Sv(@Җi~ \trfI,Ó2rs6EhPQ'LzOМcTZJbl)gkMx4)p#M" 7rN%tR ,ܢJrA(a-N 0։@IR-p.ª EqsT~ _Ɉ2%:i+AڮDZ[O*|;3I2`] %t?LcӢ1^TG]ZߠU,SvŇrٙd*\^Н_p UI>K%%(2"sJ%3D%oG hB1b8bН$ aVk&L,Vi4\df )+geHc8tgB`/K}-OѼZZz+e-FKo-FKZ-Fӳ~h7Z~h7ZƓZ~h7Zm-FKo-FKoOp6 ~CαE`mތsl >@H[9m=Gsնªq vU € >ס*97)k-ܥ=pIH\d3\=:Ny]>+j& &{H5W;u~γO_N}!':Zkw(U(>xIƃW9B&` /kҀBRJ][L( <)!W lE&hFuSbiZv1pz3j4YG.ӛ&SRL׷yk;lݢm=1}Z قLlmxÝ 9x>2SvIKa//&ǘgV!-6fJ}v"b6g@Z" 0m2*laq-|&mGg_dhy+[JOFf706A|6B&Le`(#xkĽ$UzLU-DgS!{#/(fSj4x M&nǬDS)`le]MO.=9]mv jf-1V f/yp.{Ѥv24b%=Jp[kj҆U!2@,:dĂIؠIQt}*@L5qv\NrQwv$F_gu*oøONObЭБeB;X>' #?ѬX)ݗ]oE%l|p :(@xi*Cʒ[faNЙ pV&n#ӁYF:T?Gk ;c\ϡg_f}GL߭u=ʬo;̈^W^~:?!8yOEޗڦoef _I^-kzfz dC`RfG+w)Z -x EFD ) |S B/Gwg&-TZl(fn<U`r8zM3hfȘu zuʊՔ⅕nc7s´aJzXvRK'-a- 5,!+5l\ԝI =on;-(H*&i9Fty[fF"#-ɽֻ{f}>ls']Mub̾E]XWUo{zpx͖.}8ƛ-_Z/\tS*[5IC_~Vkh>Q@yW(jN6N*~z9D[i]bJy@JgPggm=ZKx0fx^YB&6?*ơ @kՅMUwxJ÷-м~9'}z>OoN&`gr!hdc(/:$Cj^֙ .3PZg>΀A+yZ4 ŲsKs9 rpF"% Z̽*J jP;%XZqµia+ϗ+0vyU;w{;k刎օRrhKC|Cs{;6ևfZW_j&@xowxz,\ LAuA4C\缇%W0 Oà!(γlxN;+$M{3.t+ R)'Ռ*cI$1pnLXnl51;"fANbbt0,s߹%-= cӥe&Wj)ir9[&1㚎 HǦ)W"k߮bVܲԶ_i;PW|z)O/hl<݅>u*A@n> !z}EH;gtZ*9mCqퟷjQzrv\Jq{vH歝/ƒb{Di-灁1GOsQ3e )MLO?KCvYUq\tFI1q,]Ԯ_Ǵpq4zi䅝>^7O/Zqp=\7}P3[x;,˓}^ RIP<οF\eoqX@.f1>,~~<.-M>~g4ty[stz|Wgz_'dǕv'% HsҋdI| -QLBĹEo6>S~ ]X@P$ZE<: >`.AV+8ݻ[yiX؆=OߖXF0ՋZ-gWӋD_0_Gཫ)'e2{#jxg6 [Q/kQLJ1kګ?|旲ILKL-m =)A+G,FgdRR2I ̊Nj" \?F6qZo~pVŐUdѭq5{K{ħs C KR3.N~*ɥW^KOեv)o{]},6\orߵ1c`cf m1vw|8dAm[1 *%_^{5˔&tϣOv1/ø t jIM3D8[I;ky?U [K=1..yBwZixv@v!&sI%w{3dDrQ Vjd"X0%1w)E\H|7L2G٣D6^P8{x%-]_cevG* Uez"]wάUAB.pt{UҺ[6sLG󛄍;~ۆH^a:louww~1g}}Gp^Z$GiNb4܃hD*)l, fi,R8) 0Mzt:9εs1I@ he6r6B-*igY@ }ϛ%oYjqs\)Tr3hupo)弨Ь1Kv0Pgd4 !8 ke|e WypM>~Fޣ9n>e LDyB{ j&9Њ1 TT8 P.XSdY)oq.&  $ֆ,7OJڞ{ [!aVl|W|U#o_9/g"ؠ+Ai̱㪆J|3$e쭄Jizdx C% pј=l2N IܫKr:<@\w? MV!A`ʻ. %|Ml9'͛SBLr6B$Ne:']Hs ZI笑9 :̠qƽ\:'|%IWqx76FL]5|> ;/ 5y$y7+1>R#n8 ?{Ƒ_K% 7 qr%>%1H+W=|5(b3~TWUWU j $7 dSg&N`hn_q68^*愉 A#c\0NFo}=:՜䅣cKAoh*30bSLȼkKÃeNDJ(uJ:D@)aKþI!R]zBM͉5r;u%U63iAHB2.ɀB}K &|2gmR=%)?(l/+{PV,M ҐjNXl+Tف-Efzԅ2ۯbrK*ңti{(uXwBc1hfёv+R!"$ $CL* <Q}Ȟ4AJə!2!G+Pc>|8GZ&uɛ@dvAV#*,sN &Es%ÍyZhzˬϊcx_Wҫ4~~*/#Wո֤Bt1^aWpTgeVg8S5gVecMt }x/{8IUwbbb3.\Ω5DZ2tp]W0Em̿kQwr~]~TK΀+y:aV4K3ōnQΉ3:>F*kʾniֺ6/&Fژ~yrr6{p,ݟLIm9ANϦp%#Q5~4cBpi 5#q֎ #8wYFD1 "(S|8_=wޛlXrmuu[+wr!i }98U\^  sSنN įz*.&vtߔ߾>|.?O~N J)OM@LCCKv\ ;+EK^3}!ig1nmHwO^ⷽq(%֩)N7&iĽ"pv=UUuwf7cBxW.b >^>h.2 MM[^R0lYب=9ԋ$87l2ҙc>2`$HfJY\ꃒֆ9ϫk,  M?9!/9$N!BdΚ,!3 x+e1upʥ8M./k݆>N24 |)mT<s:O ?y^-Lw]O `{i~`JBKi%?XgR(^F"A=YY`L1+b9 s`յ{X\`4hAZ x嘒sL*gN-Әm>mk:^I~/\ix},QFΌJ8d!k\ @p2 A uLeP2 `"ȁR(4H2 =w$&P3SpI!n{nlbJeBme̼.} (Tb@!~L";rsƽlzBVDuc-i?j.!hetJ9&*|9PY8*í҄MEKr8%V MA Q`f$xUAch&tk~JS~f޷xk#%d1\,l8;ft=;wzx1i$Γ80Op}Euwh_E[ZԺYi]O[o9otfKǞ ͦCjYSzw ziYz^jFo7߬y'l7LӟnxV8iz5˟7ճnq].'mF"U NBtSJ93㢾uw/&qN1U%4KRVYT\_=2XDDQl 攵GkRQyrWrܕoZH긏s*_<9J xF-2:cE&1 -qT5`˝-  SYDt~ICGM_QtʚIZR$$c (D %WL>Ƴ6Ix/@g@ O]WQ:MY:Cĺ#A3C[1(A HINQ !&sNYs2J0:k!{)%g;:sHH#(@1FPI i><Ȋ >)σFT(X8:G8LJzҬϊcx_Wҫ4~~W_lGF)%wxwqwIb®2>a_~|[ NO̞UZ AwgL6!ᕿ $zWQ$䯪T;Ra θds:i8ug]W0Em̿-׮=:N/nVR g˫Nة 9LGqc[,ms⌎&\j*Yjڼkcy`v2.&szMa+KFj0t9i@-OoKv$WtnveĹ;2%i0lA@AzwmJr4$gVXtMhQދa8qlW#ɲWʖsDə3mgϮ &W:d_}cNrҤr%/ǡ4}3I%\Q|tٝſ?kTsQE<񯮱(]!vW?߽|~}Ņy?_wf:\}0 8! ೮O;tjuMͺtmkz:[ {\|U1{[n+@~㻋٤K+UI Ԟ]s!#ɣ8 MbKM9ٶn"ZW-CR_yF*K1d3,c>ZTI P,r00yM+>a!n,{G5!6$5Y،,Y`3ӖpSPgodu:G~?;Aew޷N!G;O u8!~a@3%o~%fF64R 17 2d\J42ib3{ش|":fu)1p4^hRVYIDAZ-`{LfM{ͳ[GmKR*0)#|)uot@ѷ9^|K"iTDf ڟ>낷+>>y&dRr -?>?_ÿf#W{_~0j%;.èT5I.Z2)J]"9s`7zm3}P6ڗ4\k'ҳU:}.3w޴"goc@5tVdܔ,. utPDt KMGk$IZURJ] WirMMOW|*߂յ:d f.lwW[Y[h˩AJkegoU%=J"ҐlBi,,{ڧ䁅jYUV>[uC^T]&6ohG#^fյh.wE|Fu2(PUki_` +U3έY>}K+WCYmVyo0~j{oe^|Z9jWPy߹;5d6BՁkzoCʍǨpv٫3~\~ x~(0s&*iODY=-N3UG{zEZcF tZuD%"_6Vej{mC~H`{wmH5z pfvBWo=wճfN~쾥m<9ssx@a5id V/~|guX'k;w8Uޖo6"hF%7Bٍ?IkzLdS{i[-H]`зNjD`f3wX>;}c{Afounbw_59-\mCF{'qU` 1 ; aƝӼqMtXyo4y~zG;@ {ֱѽ\YkI<_PB[>)ꠅ$J3s.`Uu\HYҗprY! X 1&d.5QvjlֱNZǞzv̈́8jJkLh% 79Ÿ?M՛}X8?#p4U߀sz:Cz n&jiM C$a*BDt'6bL$9sA˸`ѐff㞀&mt;+܆u?nWyj-z ^='ze] ŋvAJF}[e),JȆq\4F۔U|OocJ:08- x2:3<@[+k~U~;<~fD.&ˤ)۵{||(ze@;9 :)1X!6y*n&aǁG(>]Y#(?ʺ~^+Pxз (4ʌ)7>&!j (T.^P$g q p}fvgGl_۩NY`ꔗ֧Z#W8Z=r\P b ` Xf. *zhǫ5#gK̖މ]k0aĄAoDHh9^G4pj<ڵj ,:**ei9˩mdOZo*NUs#KLģ&k]PxmDQf,QVfww,p{gvCut 1Y^RZ1 ]goȽrXve'FNܠBRHXa֖ AdilRF)9 9[1\èT*Y*X3Q,ZQVDFU[ZKD8`"TQoI`t:ŕq$"d2R1^ ܁ȸV5;ijj%oc LHMBV+YK{G0HQ8(]'H%~ԂFN(As͢6ArKه݋>|C\ $J, a{8|q=ݯ.փx8\MLTDI!Z0mkh| AbZbe^*G8XӉً#d/PAY"pp)!'29+)1,.*)QJfa)g~-GL\dՙCe,8+* n5rVe;w*'p}x;# &cNΨTa ց %Ė1Y=Ye>˜7uӽ&$K_0<'[eo_19xHIH4 W^F5[&`AFfq`i?qZO4Nt~hWzOպ|O^"=ى[*Vdګ7;AĹuknup%8z'|d6lc%a\š;+s{Vr*cb.H`:fEt)W82)ښ95c=RMV}uae7}].STv}SqzM$84 ^h،ҙ ʤ$IZ'IP2l/EV f2dΦb(yMP:( 6L36A,2 dV_^Tv58m= ص"km L[F B1)]%xm!D>RiB@Fr$ !G$M&&HEYFȨ&}\5/?,OE#V=5"V>pF6hA#:$:x ` M -@G!OQx2BMHbumzk8cJ#I"p5\>HQIDI&jpc(%R*kjlֈ$ZjOzqL|ոd_(*E>A/'# G,єkUc,9MdeA8<'3cœ96w6OAs:oȭn1#dp !Ѱ#ںp, [ai?>t&usA5#4A_h%2䧼8|y?}+>{owg <_z|???VR0\E€ٔbґ)"Ġ /^A1|9I+ CΙH J2gA%ϥVXrL:F1 1NE6j BpP ;5ф)!wvjlQGJӦ, !f$:Ks=77S"zL;X3=&O\i%E$H,}c>Cb9=)NLHsSG\AA<5͓sT1SR`>X95 )[<*^`X4OS _/5 *5ǣyZ&"TDhodp*J >T1¢+U PW[ =zZp(`hQ;ږݙL X"+J9!']~j<e(I:rKm̌1{Ik$Pr&C& j_:#=6kՄ%xJ4ӴFKYqEvkq{]x[ln.Ύ^W LH醹鵋Ɇ6sj|a*jQ#4 VQL!psi%oұnGp>h8M;u)D?{ǍeJCw6/y/ 0``fA#X-)w}ҳlw:b}H^{;:umRi;:~EEd+gr^Y&q;JLC%VGC@"8:rk>v>iqoπx@)G2ljZy-He%}(4 K>ĜځF!U}b㱧=9'p\]\5]{~`WeorE#P`[2h|яx/ ]q?_|x`VOE}\M_ủen<  +̋%2 o!wlonu{Mxo^ Ï8@mޝwg+=lRqEڛO lW^!gί[ˋZ=~}RB=Z|xL%9}Fylg359 >ݻFUb<]o w6hm^yCZ1_B}gUUe9 iNtfC$練C[7|*BC&_mC*u}lhhU5m[ƂsO+<Y瓲;o߾Y## mw̾ـǾ)`9ocr^};wEt[[rڝ=iﲭ+Eܛh)9s { o^WK &GDsc ԤE) o?ݢZV耽̲SYVh BGɇe縬eg误NN>g%ͼ9!YyZ%YggO L_Hzgs/e0[ٷGOjX_a$=C?ɗwu`u9jqq]뒬l#(l?픑Wҿe8d)/凳 ׼];OelXk'#ΨQG*"KJU'IaސR:kaGtB !pFl(N(zFyp콎R(Ods/;7&vdMBSscLs vbiKl4ad2,\ZaNd3Y= ԗ$"b0QLbw^Mbw!Lb-}hx(w69XcY\`]xedYshw(:3+ O/ lHxB-ނMEw^/{yX XR$טgG+vtի)ɽ8:UGj骣Iz>t,FO?pN:ZtQ*ϑ#S+f:kLf^G Q@Wϐ N UT誣WW=gIWKӚW<{rvq~q}ڏnYF/$5kw~#3{ ׽Gjއs7V}}I,>ApEo˫C=Շd_jzݎzmՁo0F_\rԄȵv2z3rZNiCr}>ߝH_: ]w߿hiB4 L~24h֩}4si&{p Uz*t:tV]=CA?YCSX#:\'Sώ6'P>COGY JMf;4qztfJ7;0uT誣彧y-OOsgȻx:y>@qx::_[pkiKLgh' xՇofot0~E1o95׍ { wD-qV\5a|&XR } `^H>cnfGt<!<.j~ZU.y`|qsf;siXV^d ۤP#e9ŷJ֛NdS`kT`?f}$)[/ |OWkHzC MZ~ƫտʦ?6p9Z"e5[Wߒnv*(FdPTb$hT$xzg$tWJHЅ1*ܦE)Bܜ*6~1ƎPghF wBYivLX+"T{哧Pk(T2#s%fV(AuIAZoB0@_AR Rk"$mk#Tt $&AIIuǪT se{]Lf cI}cR֭Eę>bTgkOpGDW}xeTڄEnAҜqTieXD. 0; 8}Y%&ּk94YJV~|m=q.p@}VerVcmNs"H ^5kAr-PZQN^Mɇ94n "NX G_#8:Fҏk1MĀk))zSQkU ;KF }2F-UKpm*"Ux,Y2lr6UqSFVM٪\ xK-6JȤ0sc4`䒳"ec ],)j)D#;*D{ \.%QC`ʤUtȗ BѸ iX/l).!(<%nECE EG|UZ؆5uXqxoJ#Z**6һ_A2hA'C1cm̭L9 D\JQ٠+.t-c6TaCYuq[Aj+xT^!d4nZdRZJ,*P:#rJC.((Dآy,>lC?KnR2`XXq26&T$2z\ yOpPG"5jTg2Hᱤ:" %! Wћ dW0B;6QafH57._s* e CXxP ! vy4=( "*Zd7,YL ytoMI;$ XYu0w]̕]Ρ'88)ԙb]K@T`e ` ɄFB8nVj \|MEw&Q )ti7xԞe@Qߐ`T F&b.1Qv J2HO5{5k̼T} ӕ!!ѿ8ZgD,cȲFۀHow iXs6õ+|[4qXgz$y/hW$'cEeթܥ8!``ۀv&^˫ojmq?V6dz|'4 ڦ 6bbZPT8xic Ptd髒ћAhKрr2@]Xo,r+G 2A-)BJ$RDN+̫3fMZ0=Ӆa0F@_0/> ݗtdP&G͐ B@8n WFUSY_ y'vm,93|Y͌ QH]_?;<|[}%0+tmi,"z4% "Q)i}|pGL_IS`ʠvoКF nsp12v29*յcAX x +)dWXU3JDXj5YZb ^c8=,@'(YEn0Xd$ fJdRYds py Z{xw GVtE-`<Uy`*!B`Ɩv"b8 ԚKd4]oK}tԞEw֮irpl4r›ڀJn$[Q\3m,e`Qn$,T} xC=Kp0o IEzjh6 ucW9@˓y,|O۴'gu˹;(`1`0uq&Mf=*kp)]qۖUN4QG[c^kt<RV ]"PCWã!MʌТ|_ÁwäD6H#%T2ty\`uR"X蛑za)2u YZ@z|B^3r7nfX:,`?ZoD,qC(>jHjc~{Au׻(/V1 /9~Bn2pU>]jl! >1A,(z&o{Sj`Y2+*QNW8pw"A"x"oX`"LOCJ /<\p҄6t)[R%d 3&U`rt'Xۥ-` z[ `WG^dK͑ׯ_nYb/Fh:.FwA%SNӥٍ嗪`қ\&T.86O||RnJXb%RӤ?^2zj9kq$xQYgoҤs=cS0 ,߸ ##b nj  A%9JX5`@%q@HY1D!@B$" H D!@B$" H D!@B$" H D!@B$" H D!@B$" H D!@B$"^.HeT `- Zr0H 4{*XG$ Bi.(" H D!@B$" H D!@B$" H D!@B$" H D!@B$" H D!@B$" H D!@/ 5;;ǾM$ * Gi|"^; H D!@B$" H D!@B$" H D!@B$" H D!@B$" H D!@B$" H D!@zH 43֐@ }lnh T HJH D!@B$" H D!@B$" H D!@B$" H D!@B$" H D!@B$" H D!@B$A |vPBMK\? R&A ' \*kz0B\j%\*TR\1#H\5 WP}(PB(qzP}9\>F=<W,\C\iWzʋ@l+ Wr(PKžB%(^brةA= T: CyU=,kbx:dCDŽw7'wFIS4har Rq1G@m),6RR +bFJr=)]}lcw)]KRrP#.&~-I=흱D?\J ׋A I/g4R%K|Rj2:vbO/f}%<{T5PAH[}Շ:f#wkv>Ggi29fP=Vlh}ie*q('tj|5E B;㏽smz6[ Fx{+F-i,Jp+1d I'I~xϲIhȕV^z ^A:fY,6tFsW+^ Xeu&ḧ촷Xi] >9ZvME#o3jΦ%D\L*ι 84檤?r9%*cXv*]Rc}4 әFLVg@ Ivw^MXwP^Ǒ œAV]CU ^' W rth<߿w&:FXsb]޼aua:Ȣ}T?_P>E>Ee1]DՃ x5HxџNڻeC|]Wt-rW ]'hB{i죎llzG4"D z$$ RK7tNdNp^a,)LQe^%x1S WKv9dÃq#V`׭ߐ+ zӋŃG&^g2鼞p&1^^kqGK^MO˫y秊oZ+0!邙 YrGA2eLPRBVHqZiWJ!Q}ABSiLH*h&K sz5ݱR4q^W5cT*ܯ5f_39rz`)h9J b{MJ)lV6gChwD[\c{F) /h_5o{5p< Qe@M9 U)ou񒑻0p]s:_w;`7EA ӧ黇;Uvv ns6){qmGӴgskz?ԦKAwv;zŠܦp {[wƢ[7bO|Ȱ~mi@w9W ʶYjOIӹBW܃ QJ`ajoڝ7O헋2FaВ9tsMfoѾy& -f:U[`VSUіQjx%`PN3+7R`'A65{kZ͸TDN-.FtDfG罔6r&X}ѮB: 5ڸE5r Dhd7aMFW|́WWr_klsoQ&nU3mU'u0jktnk"76ʨ]Q6W]jhG@rVV_htUSzs0lV=y4 d%)n_UݕOS4lQR077{9gmx,;*oX|;TzClӜ5gݟ6jm< 2Oi^8i{;½ҡ˖d3K\fp Tvb7O%ܗKۂnx$PR%'C%sW(鉐>[;h9˦.7sTQ<{'A琲 2r5Njq-M{kli WQrJnםvƇXiSfIJV~aNg.j"VΕn4$)jCbZ:c)\lmxgu%.IVg>QM`&'GA0w^J\pȹ{Գu\3w -Mj1%hY>e '@-䚃I >e:-T.]y_Nj^.䱂T6MBt|=߮kԹa?.p>.χ}4g]n PIoPɲOyi{khϊx}8ծ_ >xυrTa"Z2dít^])JՒy9=?oRhL-fj/kdykKy&EJ ] y$"E OQjpfT(sTy-A#rU^/>+.*(:>w!qi37L;VB. v_k]*.+*ˁw4syםS70Sg*Πr6%,vEsÐ,q9GԀ!PSQB9aGVx8rq:Qjn(z!ɻDRMZM.bNnq[JI}Ucϡ)N→|LWsW D=m?~WK켻RpɌ JAowM Ts_ݵQ}UƄ~׹֧iNSU{ūɲbvsb {Nv1KN`8bR{OO@i6wlI-kmfqn6,a4 FbŇlNj6{gɆ@mͭ:u}N/8i\l%߇zŋq s.=as߷*ՔlTOm* ם;?v?Oׯ7?'{ϯ`$v- zg~߉v V8MfMs#iڶo.gmvyMh;vkc[n8A7n8TO9pZ`U v Yw´ߠJM b-B {Wbs9^?{aWdt%-]DӽE }Fc*K+a$ ?*@*_.%G+TR.@&]K40yU|a. ^,ɃqO|Z$Fk2YH#$!s`xyМuuZ96l辎x⚣lfK[z;?]v˦1u]_GaJ_6TR~qN&!r`J"+R4 x^#&[B x]]o9+cU  {dtQ$8,bKrlǒNc0jUY)Jd̖Br<Ȭ-5i`V=zKxHW[=۷7<>|>}(O'~k'Ȏ3d8UTo` d9 vaJWKUH!ڑjvOk9|n7CdJslGLLKI}@u=uv~Q{آ nY' ڦL?B+/8Kfo*pE@XF'Z*|Ic z}Ai I bs])Һ_3s|BC|3}#m}n=| [f^Ɯ(GL`ԇlLPԊ뜎P47 ؐ.K$|u]QZEBmv#zUN_w?^4S.hJLq$H 'L}@r4wJL3}^ cҭ֔Uu|wPfVvsVk~zDPG?V ˝2/B sWk5]ZįssX ^;M'QLj|z>mv2m83 ܅ U8%c1Pe. L `5Bk=ya"E D FBi[ $.:>J_լ@K2P|3EuSNGz|j0`gF$ Ȕ%F{/!WByveѮW2a1db( B|h2㱄Lmd2S2A4%E/gTF/#VZ5mfLbde_Jo±pp@#eEuն2aĎibQ{d['Uǣ\ `ƝC)%B 1(yRMhc9-#df9&Ȗ&eAFeȰQ| TtA5;l%f<}P_wR8Dl?DDlbDr:D)@*$&\IE :fM1Imcz"Y#H_,Rq&HLRL6LQ#5B J/0"~zUGEr!u6%jrgE 2y *)1U&$C[}\j/jF^@ͧq8Vc(x綹@ص冽GW#wڽ3eLSwH'M7'Ӛ|=~ʹ1J.o=A|_~+:8R2xЅ@x\l2##_+vϨ>0@S_nOowy㔜1BY j"6(z}W XY0`YCs}}N89Y uk3Ŕ?}B29-VNe4%0f:'$@V#"j_ 0p,)RI$_6 -0zY ɑD|*$2njΑŴ]F J m."ɤL]ZPAEi&̉@2Py1WPl ِzgk備+hA*zȨo"E#|amo0_jՁWݯ|a #;k@l֧%!8FU +3]]AwVnKbM voiww6a5@L$bjqw˦y=.Wv?+X]x`V[Rhu5i7]ݱf.-@U0 W[hM8V_N[2xT#Z_/W-J&|{%Hj!3?"F#BTYbU)B4#5Nቋakr뚍&,P|3leP%"^AXpJ;YΨsbT5s6:(^TNAUQj7RQm U/x@nzϸkvmB-|?W gIn(t rYCdAjcߙd-^q+arpTUR!Qi@$D6BZ׭ YB:i0,4 TcٓO5 eLR'7EF:V]Z k%r2PZ˞2g$w|Mn<ϝj7NoS]q$6Os{r~&0=iEbFL|T'~r<*5h6O ,'V[]?p|˃%# |,g|1k7{_k-(sxg-a]o J/1V`-3 o9<0/U?G[\%?Χ%/8/?-W>C}^c舯[vzu#GWkSKъ1"/IGIg:/0a'SAHtu bGڢhHDc}OOW|и(OvAu"Y)2ʘl2Z]Xm! 2ȶĠ9fЁDBz_PZ QyW ٨-%[43p. hi냯1wfhcLŠQI^j7C+9XN{sڀ 钱]! E}W*ڥȆdC7wڥY?lcGFz$ZG:2Z *ADc0 b2C Sf.Ej,z!6RMhSJǭl | Y'G&: O肯'RS1~<:inwc7tCnG?-0lq߶vO؋nM>rכ%6IOeT:]?3"̬[\O_wgښm_ڗ.b>t`0^sOhD<+9[LkqjJ-!ADS(df5'OQ|x(MQlú<#j?jc+<ԫ5/Mbً;GxF3ЃK 9q |JsXjY[b+7hS`J$}>Ypvrzzsoh< ^݄Y}1.:,kWA&{rB;XeR1-S{M~}O1(*w}b<{^$~l1ETq1Ǭ}༮ &y+λ1Rm+W[uQvu+ m{H̠*h,"B+I]amm>uDULgG7r6|忐('Gk̪¯Iv=wIVikt"#7ꬠ3@EbJHIkbG00rfZ(ڊb$+^ ~僁Nf-dB]F+ J djR D9C$wQVw_y3TKVm]OYu#w%UNv[m2ISIQ'1A00GhD6 RIYv:ۿִYSl)S*#De!Dk+DD)M6(%XfuG: CŨuQfB,۔LҊt$XN)q񊠸 \oN[y9>k.sj4^&?'Z;7P{Y/sW. ~o[N?|J"o3Ov3p2)`Ov :bH6׬%47h)A*P蝈қ4~;}3;ⰼd* XE'L.QqFn9uۥsA<=n_TLs/g>1m߯~[>o޾=/h@kbhdFwɐ_{c[n>#d0MҜwO7:w}qN?-?xw]'F1iih2=rr<;[=~/:uH 7d#F"\"-<='Ym?zv>-\;VQU;Iv+i[aQ{Ɍ'u_^{uC/;]\'a?:㻿c>w88fj"kI&"@Gpm追C^Cx1M=g\ Eskƽ.>vƝt/'0-ٝQ+F}*;;?iN| ٓ'e'BFlBG̩p_*$_AL* ̟'c;O\}Tڭe4qbx(]IZjAb|/*v4(]+Z%.Z`2aq +7?[q8[^ށJk> FPu[Z}(VIQ)2AulUan<5'ךX`*.Yg\mM9f9*4&lM6yj@wF/r5MG%Oko^{DߋrM^2XYL,<&$r<eޣ nn;\-޵\rF*dd7YX'Tʓg t<}Z 9Q],k.*Ffg|;yQ>*tD`T2J;Y7r6Տ QĵZLt bJhPIyE*՜p(=Ek@O?7~&_ѢC*I&10stJi4}@(#xoF;J&@) CrXhB"|.=hwPPM (Uè8(J$9f=7H T1PSeU&bk&,*kTGiZZ%O g"|X; ߪ.2\R梔ͯb1މRBRqdbtlG'g`lmn@X;ٰ<ح*{Nuӟ1G:~] DŽZ RYEdVW+(ho{'P#g IPm0.ےWWy|b|/r6}hY_0vgǯ\cy*% k(xB QDQ66G7܌TQ7Γ3ٙu,mBK*F;Ls[v7G="[u>أCcQ Pة/ۈJkUV `,XфEV4@N֘H\-Ĩ!Ӎܨo=v 7Ÿ-{FF[ik9z)֜mч}v&Utv=rdí?xl#{F1 n ~`]0ݣejS>0FLiHQ(ڐm=>VeWsiKI:[l1 _$dUɗɉv2)ChFG*YzF4O2kA53"0_UH$PF.5/ FE4h+mFyf⥾d([ƨ9ssr3]sդ\3:unnRX~^kdvS840tQE-|7rc[{ /PM޵qdٿz? A`)q% Iɑ{I/$.[[V{Iv2taq@D:i"H̥ LFkEA#qZxWjnw>S\g#d֜=_#7=BFg=7܃7q8 I/Y3>$/^4n:&\] 7h*<*!^Mbx?Fjl.GF%4=3#j^mw4U1M.TXfS*QRu.5-5a[ lԲBoIaM 8Ox&Pئ zJ< Uk:l ӢU9z;{hН0P0}3w|̾ IIU+Hٌ@\ 9~e3wXTJwULu69~iOno:\O4' оcڗ0&@밦a&oΟXZԱk$jlԛsdɆ<<]Mg`44m$|bk~K56 n69&u6Myin4a,v;/,XI.LeEr3,ms!h"!Xp nEˬ@JR%  Fk”`k|n\e=dW΋:8 a$Db\HՊ,l$Hp<0rkI)٘^'J/@ʃa/q<nq8Vb V+;=(d:U}Lײ`)RJ!wpt.৉n^J|@Wǥa c)- qLPNLc'} Ql=N;[B7zH\K0#4'(xAc;]'zɴߓpz5Aܻ)|z'Xu붠Xq,o>#A|eS% U%{UeayY_RX1;^ dɼtHZ STpct ˃чx>k7pZ"JkB5F؂9Lj{ !0X;ŝE !I%ЭFj*e;Q!BAsZ. qgglq$xdpV_ۜJCG0cH0Ɉ295m"q8`,u+FѴ[F#K$p) T}GE54r|[}b08hLjpyOwh`,ƶʜYʝc?^vy8kjS+$u4CT!,搪h=MyZ脫$a~;nS$li>xUU}\~qRj-(ц1[v y-Xhm 4wFm6*GT ! j`REL]yY_TxG$)%1dNF03,*͸J&ڗAT<(nX:YeՓu/X{8pc&FS ^3Xt>0Q4]} yΗAP~W9/2^/eH`2;/}w|(}} VuY}+hzm1NtpPE #BX r& sdJlg+9?vNၱXjT ʔV0b2i˒oDb (xPFSt8E$ ^dd/,f|x_!}!i*օGsAYm}q)bIB) *Hw]FM:a\!+h|c0 ĜD1b6x{OE!DZTPZ7h&(4Ӟy[s݂5XwDPVx3śGLjP[`1u.?u:oPBڄ$}' ؐ)Y{e.&:4B ]#zU/ǚ %bڙ3$Z/lu 2Sf.E] yBJLmL_j+0m_>'J՟~_|S mnJdk6ZGLTAbҿ?<gfuMb&J%].ǨX }>JYkOXG}6b!\^wff¢"p!;8=?ݬvd-Zk~M2K)RʬK%)3QL̵yA咕7bYEK*R%Wi*\.JiX.,^O%U=7%Щ$C1iTsG(y[yqۛttaG媿|Fmִ؎1:tiuZ^J/ mzGmt%,NAY ʐT"`"2_X0'l@<۷㻑䅫5ѝw1<<. OQweZw.y4EMSrL7-)Y@#9IN)6AB\902:XbR6궕:9ed>c"h* CQG@2Ijô8󪊡Cc'䭫oP t9~}:)>n%)QzȤɬ>jH̩/mY{8.2PJ63< VlWvF£<dAB8r|4)Ud?@:壢iehjgQNAHV,vTP>;X"Ub4f<ΆWU=EC)lJ"bZ |qj:2akѰ&._8ng]VZRКWOe2 4Ck!qZͨ]LG?inU\juR&arbZ}IƺO 8N*$Y(h|V(g-dvE&R6 ! \(X."²5S}/46u%G#1'!a~t=5$22B m6<ڙ5hv~5h!{Zmfo%i1_ZSnuk]NHM)DIq:|]L\a:;]Xtgs{h5P \BY8>r C˵z86{r**zkb&Qd%¿@)֘ HMkKjv]|ckJm.]3Z(Aj*"d,`ńl8L|  { o/')1RE͵RK  դ6[!P#{e?+RH&m׬W"X(7rT$)kd#q,xE9q[j%ғy_iiW ί6ӓ=Sh7ݻo#Idhve7nuO}kվbw)х͹uIht285k$B%d >4VDФֻRAf>''o\*U9Rtt9Yi!Jit, R8#c; iƾX8wy ձ]|x`_HPjA>xw? ]NgMO3d ;E^gDTA!0=ю {)Z(QDX7kWCaw6Wg/& `YV'Br};Ca8k)AηdJggi j7ӎvQ{d;'Uq|AB R2e+RDF?S)c3 ofEdX dY( b*& b s3qީ]4 CAf'" Z8"N2۽W°LrLp \RcN m,3k Ih/NouӪT$f%{bW3* :HiĹ{U#[אbʹd_\TqQ8~FDrtޠInRZ'P  KDȋG\<.fv 6we"2 dQcIpc3e?|yH?{cѯ_~( ^.CNvx1_U4Ͽ_Oԑ2`2`E2 9UNz+F6E316h8XgՍI4Z stŒgY"y՗}8NJ@ɘhd uʚt B񵳘l#ޘhYpD<"i#0|; -\$m$gEBe˾Y#4Y,(>B"G+mQ:]RJ"LJ%Ȑe8AYJ2TZFjd8( a@ْGJ/N嶴gLnu:2nxCM 0! hN, U98ZF'=eBxl~>гzFymSJ4$X0g拒$bo(@Q`y*]4z4?w䙑[}BЙ AeՔCo(bum2(:BtKV"P^Yh[&b-C%C (Rڳ;R :}ƓV(G@-alt8+LUdfZÚ "(FoHIq4FR|xR܎Of?ɭ8xn8ۧO5nWk_˗^z3Uo|us}?/JZLZi6O **3 w<)w?'>~uq);!aǚi&r_??M`+ȓJ|E/ջ!OLlKY8|?{fvv"qA_?_2 ؃ BW$]{.c:㎏]S#zlXq=Z~Ǝ6ko~QGbO| -W>cS=nW'oXfעg]k/X`\UqA1tRZ57WF?+tvR_#Ł%>>/?2 ]L2CjM~7|(fpkS3lv~|veO)?D>!/ hs1f>xI`ιN"JG}1! Hy*2(اRa#4Zx5vQw,ǫ=k/oꂓ-wK~j4\ m"Y$䨋A. jڃ5i{4b%+np$pb{ɲ}9u24AJ&(Abh,][oɭ+Eօdd7K H|kn]F3[Hn鞮K,GEsRXuRs PI9t:!GbHΗ} 'kt&uo٢夕3ల%i0ʚD$뛘0&Y 1%GeWHs`2lr W*Y"xmP h2FE) r6@4;Y)֝-ioCASvGSR1uLPA!GU¢]㦂L6  ]PnqNO덲(.tא\:D<_#qGڻBe]r& Fڭ^>XM-\|(K@$f#^KuɉMQ Xq^iAwc;c%|]E~[~ؐm<{ACVdkÁX(MaIcQ츏:1vUܱ3%u053|H l3K\=e&KeΑ1yH锍')@LQ1 )ٳ @XiW*1'J&EZS;itn ]ǀAڛo2AMyjc6Gxt!\'FߖI9u~ݫ?O?9j}]rmߘ5}fseGW}zv1}朑)~vmV.; ỏ*MH7o%|fwټ%qnW{z}TXۮe#{̪xP] &uB\윷&*22 yu\9Mb4@bTT*)1D&^}wM:bзd*c.bvmV%x,ೲI}޺~CO?y%[Ҳr_l:Nn5Ca|]8ݳU-~],~zD&CvIWJgt6 it>T{K5]I:~)L>,sP/lRk,O@6bI)f g=n{]wqy99S\~Y>gNG9kEWfDj:/[v1"Y2%yH̀wԌwyYRac(r]6o#lβk{*6M-T7QSDMKk:Y< Ƣk^|MU :qBQ)SкsUP-:Fͦb fe& W9i'M:g&r^;J:z&ػFR@}h8ODUgN堌,o [Y[s!ee1YVND%A)g\{Y59@dk 1ըڄ1X[Q!*@10EbRJf1:ҳ"=/uX(B3!P*iN`j6%V!ْJmDwl_׮ni҇7YTobٟeK2a7B  ޴̛outY2q] žpJR/`/,ލnpϤxqZ@FiPJgE0ɤ/{ 1'{嘻ooU $RUGICQlqI9_.Sn"vOMxʱ`-!< 2Cѩ,LkMЛ7Wuh}۴K[%M.{hvZC;/\:q(G9f%ܻѵ1no54Dٕ_~8vL˿uΜGz|tp8[䫹] Ƨ\^LZފ-',6uMK%2wtU̪*Jtjd,Ut|e1ѓ6N.LV ׵n޹+I~YHk؈vN٣\r n?_&3շ'gk6ZVYҶ@dMWۢ'SpGQ?bګQ ]ۀ,z|0x8~n;Pk(c-b}$ٴW#imabU*s IWƲR13bC E VG3HOlXn=_w(4O. ' Ě#*rU)19T\N=^t&;w-zkquVv`oaAuCggu_ zK~V i;ϓ#Ѹ3'An~:Θ>xSjH.hɐ7!<@:+r*"sN }F(ƥITq߂~FٌnWraӹqG>d|3z5E8?J͇2O>ϟN]Kg޸8Бu ZHsfG3 reu; l#5zPsDn@FB˝--C[ 8>>a\Vi=r?utv$l҅cr3+= R!UmRxXH=Qi' >Y!VNۚ4kXRp]rUkR kzM+K淯ҽ&"g[͚|>q(xSZ3ś^LЇU`rz\ivQ7c)D5 =!S y+@&rW1 mz~LXr}IhRhW6c& #gEtV0T1{e 2{5a~L%D_}8ϽF&zwDӶbsA=J^a|1Zwy)Zo wW륡X˩3thv[|p6 lZhBeRF]e/(#5} !3*X^F)c)ZB: OY]ڠώm$N>NߏJ'9ʷ?G@G 5}79Q;\3Q"Z N1G9QHp6@^/#*nTx¾/}^oъZ k& e3RJ^J{i>F$ҔY[$Њ#" 5,h9rվ9ےpqq z|wy&d!m5=U9Zْu.Q4J `r)s=r*'w 2sPF`֢)*r͚ 籆m~\tz爠ldk GV͎9]\4)X V Rzf\l.A/䈲ղ6Bl/jQ.w?-h@o8VIDQE4S,"(yJj!U(ZCv4 :QzjhNK&@ڧXb"իU:A.Em]^Ҭ}qڔt5筍U`YzΖz{ħT+TfbUUC gq:'NYS )ouӡK4" _ v_I*`}uL$+'1䡔V)yA P7.?in0*8\@5{AIo\XcGY֎UӢ J`0M&^ zԒfjbWe6jkOV`WhB*՘wmF_y4oW|u;spĒeW^p3? 3Ɉ2%zi+Ar8p_1ULa[dBB<"R& @OX,S-,x683@#ţx,On"8dG頤tܯSZt^rtV{6&U%hHM)؄+!ȡT6c=x&T^"Q"g>0eVpZ9#a("tSu-ֆDHRQϪ>gX`l"xܤRBD <ڮ?,EE`DsHs8@ ,YìN̈́n1r |r綂݁O|<HNjok{x *{y9|9g/ jR%4(,j)%b W(Njc(kFP.,7+Hlƞ #GͰɤ;Ƀ&&A9Z>&.9#Ê.]Amӣvn I)T'!!S!DIse m(&!j@ZG 5ChE{bMK @ օ$hTr.Fx8BD,;""/I="nTDהueXTd/E ju&Qk2$R֦7mg`T4eqyTRdDԴ+fpKFb񕖦/*UB~r2N__Zo=<2#`\`yQ'TVbG QNMLq86!wRT).W2t^uW15Hڲzaꈏa(#_8O? c-'G@?ճo+Q+ԞU&RqoרW2Zȅl?\|ŭ݊*sUCUUBi$JnUdɇ"1ɋʝ yi=RU! 񠣑 e"X9`r&iFPdtRlrL>]-T|Qwf> \rM.oXTaugk LSbz}U4K}o|=JֿE?<&ۆ458i~?7y{3^N8yWv/}eGЇvd?*Պg( e0,epaBYr֬Yohf'gsM&Y ^ȫ"AM+Ν=xA^˛E [|\lh_g6ym_~iR-><3>39?`v=MQB`/K.wz$j:A\Zuw'y 0  \er \ej Ry%ՑÕr)!\p4yfzZ=/\=Z >#*g(|4';і=- GPt&װci:L#*_'L ٝv=]6Liv宑cdzNq, ~L.P񤞍|=DgJ%j6LO׏_ AZD%)A};ݱ͛A}1jζa n7[ zh'<ӻ'IL"cUYRm kf)U̖r9"E*OTN8S%*r.T{25, <yV&q9tJt׭I6A ΡFZqFBeךZ VKu z4&v&H-=t&vrқد樇c"LWZR^#\ *GWH5X 5Jnzzpyvڜml7O)܏oZ~pp}\+͎d2N#r7,߽ˏ~=8$?b Vp=74"\t-5*͜s۱dr9S qݸ90'gߍpߜ33~Kq(ڌzirTއp{ήW]w~=|=`"דy_53ljH=h/-.^6 ﶑=E.խDrwSDh^ }YڪZ HH鍵SuI(B,0\JXK"PRn^ @N?7# *1$HmQWGICHB()PUb]vRo'-I!$DĀj!GG$p:NeI$"ѩxb& &@X -)5 J{j'B%wG2Lrɧ^yDir+WEv>SXL+|[G$Qʄ=r!" Kyhg E/cU~ru(!;J%e-RI4Yֱ<-HF9 }zM[?!޺TI{uOe<y8W4x*kmH O`!`si_VPRW=3(CRTK 6iT?W xD}8aRE *DK9 7/WE 8Dq P_4R)e)ĄxJJi{\BQc+.IqR!"e`uISomr Mmy!*/$wۼVXQy҈5dDTy(kB$^ȌWH(\H2J52l+b얇R?i:=D,?󡅟GHD#aD^G[L#f\ƌ9IDM H"euzk(!J ԍڤ4A qy)**(xT5 ASs5K["ޜD\jr:.9T.rrq[H^ r~e09RtPK{JKGKCͽ[%=9#5֍qq n~|GŴ^K[[quz'm:=D߫%2h^Mq'/Z&훈&%6wDuA Ԃwi%_15ڠVS atlb &A)G5]$bԸ!!\\0( q)`x&H܁.oCi@nz,]bn{"e5 jz{oM;\|ӀDe>(s\NE( q/fۦ`S{>1EAH3to~5M +t;;??4]LÙi{t4L] oh–m@, [vM$^w·%#2NLw_,dO,RK{9^nXu}Omw{5?xtۼ4Q ,AFU~] +Y= MY8P+H6Z>FOCțpo:^)|ŧ9qw&l'M'tjYlI 2j\UUKV) R\cEW:+2Z7ټYݎ7 "_(T˛;hXZF iow7aҍhru?M7mYiq=\qrh<| tވa'?c (=}(Xr'5x1>`c^gȆgdy|W_~}ݰ&<*K e*5oiKϚ#9Wc.~d5/2P1tI6qh~մ_c6lX qZ7&G˅Z>)7nG|?4i-W~aPK,Wֺ܍nf>F(p̆nr*Dus! (+AT:Ue)r l,f^\\.5˵ܷ?n ;υ\Fzۍk/f7`{ztѱQ~#ZYeqzX+A ,uNx+aKJApŝ$@P96(Жk넥>PB%8$ c`0,F ϯ\" d/5_Yr2%GpI_ ֞1/w>"x 9F%Jh  6IT0 0M Cǽ5:Q,/ɥaRk 1RHB *D Q , #g4!)J!(O䂣9f5rGRiJ/-oP*&)8e*W3%LhݯPI ~yBC}#+ t8#(qJ(!܋lX%a6+Ch& f"i((54Çr 8]F]}]vgzL#@A!c#&aUȟxS-NsD8@xX vk݅?Ft{2כ^7- >s{|;7opf43}1%K r9Ldc79U8Qg .N젡`C;5`I841&e\2>˾˾D\NR84@.9 щIig Ab# emQ:FQL Bm ]HRߍe:QreEU[U#gOV6OjsBvy;-n^ZyȐf}mRtj̪lw;zvh2nҝYa\Tνk< OmmI~csm]0s~l{gQڝs}]ٳ=ܾٖOwg{~nw鎎=}hzk⊹ښ/ٟml'6,ƨWIlY\FS%*']?A|ɮу; tw*U(鈐.Y1>U@gLN9wR\{xr.ENN>&Ad4L0ǃR\2[$bԸ!-L%ʬb X#5(O @"^=u_]g~%Wxro/ fM3tPmWV|HS~fw7W̓#O2هx&yH\.\%)_RËHI(P$#uṠt7jmv^rI5(2`"hd Q:i&ԴTS*։d.[rΪdKNu%@ggOgq6B*ڮN9Kv UGV:qxO1q%Wg#R3Me&xˉ\˯GTIy4u kDFr#fÍ)'yy{=&!ܓ(Hr>ct9HS<0SAɧFVmO 4u),Y[Y9 Dn/YxJ3m'2F'izIC:OZ'2KճӍ?;ObGgb4F Vblnjv9`nF9"ڃICmxYgnnZko>ݢT݈nƶlCݕ"K[kq}Q)/ؒ2r~vR7v?Gc{mq{@%&0Zl2p,y\)3T6j߭I.oΠ\ݸ0sP0"#3 -A͌N*Fd-&"+n?swn{MFAΑ̈^(EvFi"âhJ+ˉ\Ȝ"2#;8 ڴo…_Ҹkg)E%&Z!1 g~<jt2@r1#0ku s\y] "TM8bT/^^?N}ubvJa<a3L]SH (јQzzm(W`=w]bVy&M.'bNnF xC49Wc+G#NJߪ (f~5}_^~TD(kц]+%ͼk+Z'hHͨ7}ԧZ?&W~j.M?8Y4nST91koӳ؂ۅH5~7Qbҩ+ZⓂjwW5#V687Ydw$1 ~,XVm6У6{@lU:U}N_sҨLFJÒP^}5J0WEm#^ 'Q q륚5/UM^L.WwqFRG~ND稵A1ȬfFfBx 40.n=VtFi|uD]-8Ck<g`GĞu8?]);;/seKMRbkSzI7?ͺnC- NN&\NrU>W{L"tS1K!0Oc$j=eh"4ъ6dylfVYE.]u,t9;G)֎?#͟z̉6NDׯC8 5oKc,U׋ 9,)>U%ߌA5zLM4An ={3jj's`e"<]hn{1ZxhehcsZOpa]o/c#Pl7l޻ծO.kYE{tȰXyL߽ cW#)&g/6,=k?V#/eUŠR'rrꠅ$%\s.`5BB)V8+8b5U ~Gّce r0,M>'HJ[O괍9׉c10[{m>:zQیuvta&jiM C$#Q)T,OmĘHs3Dq%ZrL턙ns 0NOX&-==`u= \Kdk:lQ Vۍgq( &+P h=PLR$l%if®JV]=\vHMè4J]I۩L=G8 uEky8ꪐšBJuCu%8ںZ` jyidmyɎل 5 əF_oou9On43憺Ek,Xt!X.Z}(jՂ.TwA+"X+{0ꪐk &j {uUSWߥRu5))8opxQyU5)+wk2HF?saiyP/tlҘ>닾ZZ)Ps^;Fm[=iRJA1&[[қjL-II`"EN˺YͅlZ q*{ž,؋ORxъ;?jqpQG?0Y1 1NE}4r2*4IBI'=מI쐻 ςFJF__BR_\Adp.M ff~&FAp[Tu׏XV|LiM6vwG]"贏A̘ ,$3'0#5%-}B9lԺ:(5"O!*'Idei"dINZvȹߺ,:tx+ m1 `R.⃀ATs#KLģ&k@~SlC H6`ɂ<Ink;sIw {ݲf<3JO 0N`r#\s/D(F '*geZPqb j/.E&imD&%m вFΆr:i(k. euQ0R59eEБȵj H^t"Ц|~)?U]?wLNg2!ZD2"}LbL@dwVkV;?n@Z+1qV&S$Ma Riч\#V(rrмS(k'^j`S y:i2SLr+ .ǽt!I}?Lsȝ! j">۲jVn %2GSdgVDHKV%w%rZd'۾HGkH]X@^W? EԖ^D-%e%(٤4M Թax~>L|ImpɊ7}?$#or"1f'\L+{UoW_WDԱI۹8@}q݆ND_"6:@;X8D'ɜdn6]L5:)͗oNg/.[M6^\"ZlO'?j'ɐdAoZuT.ŨT̈́%B[NcP> 2Ĵ :-%> Q-qQEq3 CPVe\J2G.hsL~ x j8HD +a$Ϥ=rD1qYVg9r춷j OϭU{9=@;ϣO_4S6ɘ3*(&!:d}2>1 ڣnU9s7"W%d f$ ېd6@K^~ɉCBOB02zMb,AA/^7p*z<2z -o,~do*URT[E<9nجY uJw8V;Z٠L\002[s9 )ۤXI6b.RF?)cb.H`:fEt)W82)5ckܯ۳UZӅqcua`:]U]8m|U, hMѷQGrC~o4~t&2i`95D$^G s+JEa5ɝM޳6r#WJߏd n/l.e3i;%E=,_l%YLfEXUGR' YBIbTH'K!#(āBDNUXscد]ٰ+\[\`F(:iF RkycZCۜZ*IHV~ )h bZA9AQN>r(@}J}s͟cW8b68#Y [eApZb(SmR+f5JO+Ld &@HByuz0BB2Ѝg($^2:RD @4Ij$f.U{#~:⼨KlXr(_$"bZ8i=2c rR1sn%ps˺a^cŽʏ– ZsopU{vtG@FQׇPx瘲WTV>j %p)~-euPR⾬WXV))ǟ71vFax_i61N2DO}N{r/dcJk5Όm k9Lg q` TxJ kހ[7j08RJB?х^b8]4?VLy133{*oD#*,^+iI`DJθQ4##$\^1|=SpB<9L :F$1qԉ`1Bi &tXF>h^݌alM1-ҖAw{$Ŝ6VhW Dxm8RLUE8:hxD|QL.FwbA,6a˰8WFb"(å*r AeutF)Q.bCQgC`4Ը[%M^߇ĤJϻ(Pָ<ABΖmwo)j$(e y^C_ӜfXAZhhH۷ax=jS~cqRyʀL $⬫/$zKq}R?喛ǚ|;:\4~yݿП8:'TQM>yH=sd__bQQhT7$GŒ`qrJ=}Х|'쒭ȓ͝h~x$RTha Ѷ,#P׵-նֻ{f>9L.njnnxd4MkխvegGu+4QUފd-5zGN^5ܛ5+VmfK K2Q$ȵ42ZfRJo,QS2[ք)ϱgs9 硩p™ɏ!/Wx]́ !RB:Ud  R+9h}s*Vil^0ųSơoeG*8 ;ϫ^qʝ{RJ0Aie2pD0SW|1{L2C )B ơw~!".Y+[Yu PeA܉ʂ8.;P#-jQkDŽYAsi#SUFCOa*AiU"bK(F[R.TTf/h̝28DH|KٿWFT[]#:Op.>f~nO3yz KD)d4i{FbH0Ɉ295m"q8`,u+@G4o=G$p) -.}mGڨ;5wعzfm5WDU-f}< J(+ޤÓ};%,~Vv "Rο#mc/l׎Mg~;2WqkY&V-j-g?m_1N4<ˢ9 +(G5XE$"_hY"MC*c_jUzt Bi~HR0,D]-,b1F: lp6HF7g)ܓb1VVcMD0X\%UUV$#~:ڥ#:9 ~ \9m뫽u|9|%7goEdbm.審[>sNQ搢vPkVjZEr*RMJjR ;,$9*i>l ?>OwC| =aJ3N}D1WvHeYMX2U񻓧&!|@ͱZB K1q7eAr)Hm ejo'wКoyRlo]oi-ZLb+eaMv[BnÌoo;95|k}xqXv4]i5}f1ClPP<-,% ͣ1U ~6^pKU 63CJ!XrXC8äT(dB3"#Op%YRCAdYDK՞1h깃3\DڔV4(i:2N-(H)56UUo~ pz6FKN&ϗfvHwL@B.^݅9?XerJE[^+M%NʟNO3:R iK|t_J\= 8XF8؁p޻J=:ySS𠼋! H@ 6QhJ޶[]-rY4W7nu'AyԕIʯߕM͒Ÿ٨`Zo׳boaw?8@Z}Sˌ`vr=[EEϓzKVs^e"-רS r$H1I^ա$d}Lbҭ{㽍LɈ7r=%0@F3ڢd&#Q !x0aKe4hj4/du68{FU0G-f'CKDVUpݕ5|fZ;pGpDY`g`ZgRecvG)ד.qIwVЪ bj޶;}w*`.Vr iOE:ca惲;vmUxZnLJdb%5 m蹒Owt U Z%j6oڞy+l7/<|ϐ<ɲ2%*P{WƝ|^eƂ5|6twE= " &)4JpQ^wu^w.j)*e2A"JIR˅2J"@хXIYuwˆ^FcD2Y+냉KM-a$EV 1̺{6pgָb:y\Iɽb9'nZb^Wm_eײݙs 8iĖȜSJa G12+Qj-(ц1>o>Ny-Xhm 4wFm6*GT ! jC&w>\r@I}mI^-}6qNQzEYf)QxNKbH$4/% 6 ڤ0E [Xca" WY{m }zP޷, Os#8P":weMq$I[ۚmBc65Q0KT h %_dARf JoAZ.S!N"EQН.mT%rŘC, mbF%"O)dzҥkΦpݽ k=0%m\v;„l=_Kx:O[g53Aޠ<#?jn7D0ؼOА)!d ( &`SQ! . d7k 妽tI/!e ΠQ$ZM:ʬ&Ht)@Z ;]R pR+H3U7fN,#Qf}z~()xf۶{3A=m>aM5k%sޤ|~91`xYh0$Ux{V#HDVF֐,t1ձ9Q;: ^׉o }HGo >y网 L7ob>4ZWӑL Kˆ $mu÷j2 OH)E77Q|x㝣4+_㾰KߗS͍S Hqh Sgvّ߻킎4XdBABeNUP&J%P{X67 'KFn\" ~I ꄽ}1 ˬB藒VCOXVg ',YLk\l|u|fJշD'JjM2@p McdDT-F̊A۫2]}ga4F {\I,$rɨB:Ue xgu]M%"ʔ\LQIxө=NR"ipq<6ee1ɡ, EqMgG6= cć㖘_06RZL-ĵ.rxDؐ>ԚrlGf:.w2.BB:R5i]ʤ%`:l,jES@Mz[ b|&T;l{?'y,ڦ%Aty #IDTVzܷǖpgźrݣo#[R(ao+EULP{/$0 32=? ׂa=?k IC! HۨHZ$ Y0]LلJG:>Hb6CDU@C d!R"*RJSD6''L6Hc TPƬ TX덣UsUkUib ph8* y?_-鼣Jhx'8Z:(3sm5ZVhy$)?9i *IU,TPRH-ѥ8z,-h2/|"0UAR"m0. J*M_x߆5ޟV ʣiJnS JCKmѾB''5ڀ6[5Qܩ9Hk_,*|uz niz0m^-k?^>x Ov\᨜ߟMV.)G藏׭ bs/|IjX;#]FfpL,oЦ<`ǣ/܌9^oXoQ<|F]Vw&gl|6j~ q,`<5{۽N:L3l8>|?Tz{*?}_oN#sxۢGk==jJ=b_ah<|haU6|qrkƽ|,ma\in?_X}y3!څo&K u=J46=d8YvgVF8yC"WWbGSuXm5Q+F6&IlW!#)eA@D?S( +.]B`BY1*צZ0|= Bx"ܮ<8-És(9D,xƛbD,pBIpZGM$_|$uV=Cg4qK v2% =?):Hœc9(΢ X{>F8}OU`;?/kQޣ?tZ"Nƣ|ωgiTdQNxi.i9 >bMdE*c_|#J 5iB NKǂ=&?ze׫Dp"hJ)ʐ 'Y3dR &eىbKVWU;Kmč_c:L\y[0v6 ͠m00=zcg'GvNCz?>OBsBՅOf찈Fkh2NQJ U^MHzW;\K|~S5d 2v WuނmßG괹+e䍃[3xPPd76D@r! ,xyB(8?u*WV.TjxA3~YXS)Rb6 "HaR*( &)pB_,lYr@%0 ƒ5*yC9)uȮ]:wl ߹;#]cH ,ExMzhN&e :2"jݻtn|'倂qԅ\ k*DY\56F(N|G;.@w6fAՂPހA>Ĉ}DIx]!*rbtt|(4@)GGTQ*: !P̬/;?ctƮ4]grZ_VgR kx| .|19CyTZ&zdT@`kt`!nnͪ<af5TJ |5ݖT㘂V\c` B|>ʕp`u` 0E΄Wi:[Mk:^MMX"Oo[W<:/fUy~ЁIXڥ %,H-RK/?l,|@UtWU'ܹM%D) %)0wED!I+Pdnԭpr)5t`s kq-۱]˱P e`]Pc|tvU} ՗`?i4lpD?ͨz$s<5x:\,W72MV7ϊ glvK3.{O޵[:{$7Nm2rD\RB~#R1$4ܺkfzsɖNmTͻv-n̨un7m~oo|f١祖jn2g|v}}?^'M&pD &^ve^v?PV 4hU|8Evu0֓.B%5xa!ש@lΏhCTtrA"ZE2yOFቊeΦ~(6W_^9l.v{9]Ca|L }>YOlOgMF2MRr2*OY"#2"[cbVE% ZT&*Jf@iJfI.$e 3H,CQImLP(]Mg ^^9{YxbW/9nWSQr\K$Y*F'M1E:#^%R5~5ز Mc0JQIj&IYp@R>jXfDh1`P.U[JY(rMYtV)"ʔ\LQIx)l5/Sk>GZeXB:R2i ![!Y-P}ʧURUnK鶱K}Շu'%%IRQDXV'0 320YȿOkvp5ZS X mZ #=S$*GjA(B!xS6!C"=2bԘ3D@JDQJiJ2ȣQ&$޵5·ڑ4ʃxTeRÉK46EjEJTi /.Cdxku=3n4*ںc+-%4<~.|W٪a2RJl앖^_M;C EawWea?8Hos,8m; XQG?@JgZ$n(&8);h4NI]Lg'iűwY* #c)Ĝ 36-.'b%n6ģEAi<[zhsD3o^:^kրΖE,ry۷$Zխi8#]ѨroBk8{0?6DȻ70%֬9_<~\'AO gnd稂muy]v+wJ6RV>?Oΰ\|<(q`2{[5mh?-[pxzLomѿwG߾?^ӌzR"((oZ[ Vn3y׋Ÿ jrǸǨ+h!ր> ~9z2.۶+ej$xXWq埐d'*hT~x[CYp /m]5ڮz~?[GV4cI e1d Ig J;BtR'OX7PqP KkZ}+@<8Dr4J9RY .3)  0&Fn5*k:4;?}v8C5g`[G zsver~hߐǬ켌W7K~Vi;ϓ#+G roWI8om#%SMH7AK%J螷|Bd1d9GOʠI&LFUߓfqI7zomw7̾^wlH!NߔQ?,~9߯յt{'Ǟ=;ژhֲ޸3" dj 4;M) 9ethSV [W߿roj׷sAlڹO=z.eKXR5r_zi.$ap@ȧOgeB6~D09 a4K$?PЭGB| 5hdA#|G(w]bwV?V^C摗!:pC֠74B+Ƭ d0e2GY\rP OeBѹ,2k) $b!f^vj+B+v[x͔4tSz9^!zLj|7˼Ehsr*cR&PX4PE笧OD7,rf4iO}<_ ɚTm_!v-YhB6^d_֑?ͦnz 2:5D `,8a|)3:4㻑dy5ݥ=47ޅy&F9"h"e6XtB/mhdP9= G[#%OWZMɅdc̑h"Iǹ JNƱvsjϫ R=7^8"/Ü:k Qqo C;U}2)~SFn 2D:j#&"̓!8Z[n5mn:;`NTΐ,͆* 4 W19n'?NeirC (YU;,{0v"'n BR{I8a1FYYl)uHvVM-l [ܠtxWJ.5:F26CNY10rۺեOHA&P.^/z ~n'|Wh&͵dII+'e<4c$cxjUZ8~R'ݨVHK0:o$ HAbF}[tS]K rwR$T:JdBv:DSR=29k%,xsnRʥ'|IeH)!(!8.e"r>;tYWg}:!{VN%YN<(,9xdAH́ 2(M9TBmrI2s:pM&-ϊb9&H+VrT ư<5Z-q}1m$II.k!)rIIEIBK,-R*#b5qG/lj^u \6 ڥXg5-y(.ʸ{\qqct?Y<ɖ$oPE[vr.cym#'(; Hǹr${p=.>.O-<|.'UVXV,sV w~|+" O֮ͷ~xSճ'f-~m}G}+$=f-Mt7ޣnL WeCN:EU6_O9l2~<,$w!*I^ T8[c+,iΌHZ]ΓEfq{;)AK C{8a |y=x+w^}>y+x$_Q) s`y GgI1$Qze[:)O6b2DOEwc ?0p "r$raI9IӨu \m]Ȓ'`4(M,(sD]wV^nH&"^r Vg~!]BC@ӑGGJ/x_r v/S[":,pS㋎_4| TlR(TDr{e,8$^:d1aހ쁧LfDZN!OQG!rrHIG,'g92Ta.J=TAB=\_kc>Q H!S&X>_G dLG(,sjVXXSݖ  '7ڞ0/ X7Y/b.a;4{0ldDRF w1ʄ`%\Rp2損Aa1s,`A[jki|4۽U-r2ft_?v jJИ?͎bGWCM+BEɸk{ -ba-fWNzxWY瞏 ̷YqrxR`Yfo2_T^ئo'Ώ#Ѿr?>ۘ~8\W`̽i48^\WubO4.Y/Yvh4 6 o:[5oz[A )-7&qqǮ$EoJV"N3?gi&xLr#I_v]oCV%``!Xr]"%RȡDc8rNWTU?U]/=`^3RYOwWw2{:!_bُW#7=r~o .h׮n߱bL¾c}ᱺ&E3z,َ̈Ih {(j͹~:6={v7)\ mܢnn1IM4a5w=$=Fyb=F61bt0x.gJ2 k3gKYʄwK%*͵ƺsZQzC~WU+xc ezOj춋 ^WpR-C$G'uBr[υ=)wxenLVUQdYnW*9( ] KMt ۻ/G_`/z R1,)tiO@-[kK_b6BH(.`[.&X |l๐YT)MC,E%rń"gZ/ia}VK/ѾP~j`ǟCp=3>[ZL Z@LW Jaz=1xuuhV 1Tg,J)Kc7WaLқjMdwdjxO'Dr:Jm[&7 tʌOeN#T 0W Yq-k[^G g3bW^(Q!(^ٌtl cU\ v0B.Z(ԢwBRAZ] Dy@l+"W2{(PB%N]} nh`U!PUVþBՕD$z7dїzwu0g۩5;Zi V|t[7vt%k +) >+/ (ht`BnTrY Vn +D*Wmv8n79w{ 3ob;*߈5%.)q9|,n%DΌ)DB[#W(bMBJBH-LVڌ2f2e,2^\+Dniky]'(Jl}S^i|t e2m`xN@  (QǗ*Uw% 6FnuNpg:{^nDnRV,w5ڝ/6: oG^rsrq+%bs3+yPS(y9hޝ,0;ķv(PM)6lZMm R@W赨 0`k#v۩dïvߙތVH0"*:JQsXtVQie1.s2Ȍl<)4\j .%E76w/L*&.-y\ki3̴5ze#aW(AtcRf91imMC7 *iH!3: l,`$D7I Y$@zD|/{< UA8I睰.Zr` )Edb Q9#0d @0~tŎiE%‘҅?t48JjM5%U9X.0K*/u`ڤ3r'Z~8mꠞJ:@.{+wj%RQWߚZaQ%$P)^Zwѐߝ:BPjR0qMmixz&imoaNݽŵF.L/|s=>k>8fw=#}ydmY9Hop9z\{Of{H  .V W38g7$1 .kD+XV◃OӅ݌ٿ$4/GNrըJQq%/u͹Q]^Q/zP 잇*&&{uFRGߝ|7ק|9=ЎZq %k% zkpgO~zЪi:C6g, W{W|ƭ~m o(DN./JJb}QZnvIKT@hOWlY<﷨L XߖLK]~Ȇ$,ȥ/>ZD` 8WYZ->2PtB*˘UR. u3 HOla Ɇg: N|Zt̚mFid,da3ӖZ0.0ݙN+g:4ڒ}WF<O,^;3A13l$ާG;&]x.lyvdL_tyJ_6TRxv2!y`J|2 %%"lSniNnr)[Ϭ\t3ٺ 'MƁ.qV@:8dIK-U;'FcM dA ڨ" ٨hM(y[I~T!S*z+|nw'wOS6zbьWn:z&t ݡu[Am!!Ϫ"#;u([! u%)Ur}u)ۅQՇhͶCZτ3?Tj\8lTsZ4| ڨ@#b1_yi2*0]m 'r>$Χ3k RA-/}RFAs%3IDfR\5BBa|);jDy5S|VxFgWb3oy&ca*$NsB- Vv ZBM'9Ŗz®Xh|^ڊ3 hB+4d߳W]Q%B,b +sY]S2hp!EdjDH!*'IB$1K@{HApmEöeѡ ie8-RHH[*T`YR(Ě! J :!Y9:mMrU/@[e9 xfh7X: dL wY@eހ|@֪9\pe#'0=WrM2BYd&Q'-Yk)g÷-Qd*Ys}d(+"#hUi*cT^; n#མZy9zefkEC x93YCyҶz+"$[I̐ X9T|e8%c2=Q@ c,3,!Ƃ2%1je#H%7ޛ'`hGrvVW;eZA/V.p:BԤi4}/!eE $!a{8 X1_ `Ac8Og̗|U%)SYϡQL&DꡘL䔖X1ݸ}m]5:&8+iRR 3LdF$y*FU\*ڵ)э%pL/ur ϕJ\dm楲^̈́h8{fjG+"2?OFUNL*:B8ǃcqj!eLļAbeeθ:E]覤鋒aǤY cɾ%s1X :ك[.KӲX^V'SYyA<@ɴ{}ۣhL[OAH'7%7TY&dyX!q^Ӊ|b8^`u^*; 7 YDf 6 @:R8a&7 (]h1]B= fva԰#j}sB}Zf\;dR,1@\֢,`'\sFb Y<A! 4k-¯oi/߉,EZ*S\ڼoftɿypo[W^o+rjfjn=ie/BHGFa.E ./K"yZ?'Qj7"u.Z/:tQ6W^?<$eׂbpq~{E[Ҷo\.֯~+/Yasn͍oumż+u~(`Sqqvn}\a۞H-WXq-A줿rԗqQV'X+I9hP}_5xK-[Km\Y=C&ߓN[$ֻN׼xO<̨G tPjp}1j)o2{aBjˍ6da{G )|Ŵ6sLs3h ]Bg"%wW sVZ tI7ʡ6iRj\>A+'X;߮`Ɣ2MJnӚ̔ .Rѳvʾ#uwo`__}ҁ6LC"+=}ӻqmɓGf]#'چͦsj] 4fg~nNlnyQ.'}g}:n0O?κey3dʎh׫uh]-.7_%嬇OyڨZlJjzN%6XqQ6>y(MWw6$ू!`LSֆdm [TQ*>RU Nȣs,6Q5*x$teseKأ+q<\GKd>*ɵJM< ׂ%42}VK}"Z<ɮ?oWKX-i^(uI=7^~>kv#lc0-Ǿb&z2ڵU+XӜ<5Vy(A @ViKa7h&x:6 Y=: Gd8Fa4P GS872K:$ gxhv-ɓvvUg$I75B5Bz r3bf&&:eD %..!2'ً5r 4.)֫v`4y\\FIS@ fԜa)hmAg2 tǬtlD:L,}ٺU/q!Ch nK-4-rҳˁnN]|W4 M[S2?n(ogߦNn}%jw.sjܝ{\pG@~C{͂ ivǥYxK/52vε7Ja,%rx ,~L/vtIx&s/*ju]Ķ쪗۳ё֓# =[}pa/@Cmkl]._'u/VuH%O_fPﰘj'_8d5ZYUɫPr`XjPjB# 5j RdKzKY+m=&@s,Rʛ/}g/l2[anu%i!ʿx7O09_LGDlF{XPtƷt}vՖŔkG8/RRA8\;9}mr*x] 9FFCN1-xc6M[W۾@scykb G1>z `a*w@X/j\0#&@خUϙb+̥)W$l)"VWM:E\ f / W$bpEr)W27v\Ja&\"p?> ,LjB3og_wC.+LK92o߾~=y'X8O6Ι;jΜj}{|K>+(-$WR0Mj-;IOӀ%y$XbpEr +RF?FM:E\)nȂp!,W$W+T+qE*ńSĕrSN+t)icTv  L1\'KU0Tӌ) \IA, %N+T+=??H'+{*N:k2|3c nξnol~)1JzZ!TZ ֕Wt_开*Y6~j/~zFv>/p M|g|GU_bUZڧJsQcj*Ud ˺o;'j֣%~:oMNX隰eJ?US5/X~Tx;GC&_-|C"nN2jMsfTZǩ~K)%VEC²HeWgo s=蛛K[fj!mvaLpi5;}&|UZnUR6=̾+YD=*mnXGwc\^?n2 A?~9<kmh[&mLdm JrVP jlMKmNgU (W?-uQ0m5W[Bl(5ݿrq?J|vS_ 9~f Z8K'F{p%ty@]`t \0\tZuJ~@fiO$3/DnmA+H*fNԚo"a AseG~.Np~ f O?]^m3B Lod*)e]w#.9` xj Q$\1:ZEbm7ew0ܮ|~0T(OWV_r%k~4Xj͈#kW>%<ٺc~cH`:^K_;Նw\Wtnt>Z퇎[VՁ0ꄖZè ȡ0IF;Ew <[اy\5)9ݨ֌9_4[[N;1pN˵=:H}ܷBrG da'G&t̵8 weEɛƲKlTO5iEH b-5FPԂTnا3F U Я˙!O+R&\"8b+\\(WVTiqPwE)W$e1z?HSĕQbW(X@9ʕZ+RFB'&\rtABX1B+R+qE*[N:\9-gWNj SL }U*7Ů\AǪ3S6RÑqK.#ǮzH^*w=pvznж \ѕ/Pl(\\eJB W/+%83\~Υ;q;P{ۀ|!4 /'WiR1M*U"1pE-+WԪRpEj;HNW`%H lwHPc=qT+ʉّ\ uWMSĕK^Px+zʕ+RF+R)Sĕq8X1+Y HԎ߻"wu+)vEU\ZmƎ+R:k9  VNbǒ v WJuzuƜ{᪗`^O#']%Նb\R=p&\ZT `*W$WRpEjǏ+23qA tI19gRۼG֌h-b_Z{F~Ōxr~ή1vLAQ6M+f5fAe;]UjֶNjw86hhk΋ -\J -ZZ O -sH( .-W'+RiH)JqFp]/bpEr.WqE*puҜr+X1̤V~Tpu2 (ɻBbpEry1 JNWVpQH * 6v\J9-:>E\9"G+lˉ]\ŋ J3 \UϘ0㪟`~rݑqKbG0O%a]KƸ(W$w˹(Wq*Ek=tp%8:Wz[TbK3Áo#RpaQ I`*nǎiRZt4at0-j?Ə+,W$0v\J'\ @ڃd1(Pr+\)"W1;q  PN̎bHtLN:A\iM 6̈rfDQ/CWRN:E\0c8` ;HrNWV1 \`\ܣo>VTZ6q唰 6\\'HRc4b)H $\i^hpr+WF"fNW>z`gʪՉo#jG+4y׳/./)Hc֚SoCxH8տQпΛmtF;ڛAIEH̹ ߯?OՆΦ4·%i* ח1 [Ʈ_Mt3w/;4UBӚ(*UU)/w[.~vVg?vVɋUgMzuuϛw-*jgﰥܯ=f?).k5ƚ^u Vv3(b@<|`>w( p@j;zh̞rKe.PO{+h=-~?zm&'ߡ> 8cʦ_J.y6ĵ+EuMʲ:Zjz#k y_~C-ڻ;> usW]\q)YGڠn0v^IRk=SrYEvzq5p$6 %a&Pu!J%&aQϱT\ T/LR"sXqAc]X^jnj ŀy"6g9jMJ>%Pa-w)E%Yǀ8@\}ީZZjhtNB_C7WWѥdK9+b-6Y1Ι(J(+Y'aDӒh>MB 3VcBsf1 a:cM|9,YC=&MCh@/7dG-]-u*SbDk /x]g*$ Ö%[$mmV4*%}ٚ\6zhV`6e2kwiBC–Rg嘃}~L>O̬>pĜ根AXFJ+m\I 1OZ(S ܰTg"S6JqKؕ#6YFek/֨=$l+ulJpgap\e,]trxoS[muvZlc&Jtjإ,J0_Z P`[KI(gu4D"lK =4A -hL'fj}&\%6hELQ,C `Вg]ޑqMLػĚg&/k=еrrX@;@R)L} C)@/%xBy`fka;sHW};-f*I3XDCB/mUZG*J|E`^$y-yȀchض2NzXl:( ¢uI.kh@Y:XH:!eH4t3FL씆YaՆm(-Z@*3ѧx"Jz>{&\fH6+4\lO!(Lh"о{$O@8dC ;gZ{ۺ_|? -헹-|j#NҋYd7(K vy9<+f@57.WRZ! c¼7jbȇ4VH,@:ZR_7UնC: xDw.U[ct#ffaK-0&lJ%80)8ЙVNecX"oFQr V'J[e@SќQ3DpaXUbK@+ho ~"RDbkj{PPlBxFd0ZQ $WEbHc`"@S[4,pZxZ댉pȀg:v/66eoM+F%̷<k1 D%eA; 'A<ƟaZqmtirXyצ5p>nG+.AU015 ZǁmZI|4x #.*҇Mhf%^E Hrc005n8&8dg~cR (ЃcF<@ $9-+d^0@T. U e:XeGƁ6S@^\@$6X4_ 5<oĭ |tX:;(*@DU;͊qn[&sJ $'a Oow[~>Lh6yЎ2vp d`: (N6G}ƪڜ׃bF4z,pZnkT8t-I \9vil#: gh ` e@Ao 2Zdp?`,p0",0wϢb!t",U@k6֜?倱͐ڲh0M`QIRhW jPrVYA܋,ʺX@K039l@E}׌I'y*0t.Yu0qdbܕ0ex6-k3 %"h 覫2h&x ѹa0`ء4eFEֺYCm58EMj ޟATXD 4.1#dnІEB?zV"5R&~V 58֠Yr@qQښI<(0SLGjJҸh µ-k$Gv\M>›Z#@KvGVspk o,~33#,+A-t1E zAu!֯F`JnÈr|d8rOVS:(0q˶ %&I%n%D`~V;͠>rI"VŪR盉`ebBv"JI-bJx$"Ȱ?tbɡ.klץ5HxXD3 P GVi]ōrp;D. Ε;m֞ʏ ~u=r}`bAl19Ւ_۫^`[ۛ"F?NWbn4Kb r:ZoNa2i;ߗx6zk?t\MCZe~-(l<_pa_3m[b=gVr.ͧi%m'~,eQǾøk dPB3SOT޳u z9J ߶%gڒQ#@֓@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H ̕@3@HUj+2剕@g.Ozm5mח ~zRb~fD #$E4f/Gsz1%X.Y ҭNp^Lj:R߸֬tڷJ]-SskYrք᳧o|W50F /Lѯh>[wxS!y_[~->뭕ߒk_,Ǔw"Ls}1)~p1G)(5"*lRN?>V9 D{ߚ7P1ͶSwti\4|~"=PjoYG$.h~L{L{DŽfkV*vQ)nD竒.'d=(J/MQ$ً TƒoY??ljMV=՚f6Z-8o-{4x#8+U+ 9f\J]"t`Z310?0۾ݏ+m9w3lt.:RsL&nLM̉umgp79 '$dOl- !&" o1 <{ͯ |6 *_ts,rsMp^2$a{}w4Ji8xlm_`|jE%gpᐂ;]ͯak5~{9<( 5{s {cbf阧17?L&0M}B 4K^E;M`~b@Dsc{طvexrg+/"xsQ̙/%sQkP?!Ò!fUc`鸋i7v0:Qlqk{܍(p3eسk26\;^g㋞/2j{tF4m1Ko˲YםJ}aOa7:/ti렿$ˋܸ\לo^z&slVx֮^(C̍qsZ<3uڜqxruݝ̬8κt :#S:Vw'Li 4JQ8jm;u]Lڤ1"Y52D+ur!JfΑ}Cj~ sǗk#>v_fwMr.ggpDOl>m7TҚJ;VTX.5zoAy<5G̹{Y_8na߽Es)VTr? fDR9}(f8?Dl.ޜO`0Sa.kYz#Gbϐ=<8 ]P Ǵ[SOLoKmډ;"8w" =k@99wcODF~Nݟ?6u`Md5!ɗ1!'zSA &PAEǜpDp棝͇{(n͢>}_z!ǻ# wgAOouy" |p sn[W=N/y(I9hNE+4SEc{l{Ttlj6y>=^5Fd㣏f-jpƊZMiS5XmSvG,A'Ij'],&RY4bpA./BqM5{܃s$gDg4+ [5K DsmX}a+H WwkIfoN#=mſ!9`Xbn,c_dcI<}+$[v,[i[v:lլfWUZ9E$4(E@R Dq쩂!P"+d1I0 ([dne:R|(`3*C:gvhtUG]|w@$Tu g3:NٜC< $# A)6y\%ih2q9}1Ly 7H磓+eiy^ y??>_?ԁ`jc_Kpx~4_<)~!/tzpiFgSڽ]yգ5/3?'ޞޭf_ST b㗣I}8\-˹] &g>br4 #IG摎7 68ׇYg_Yh3M~XWd|GG['rJ>!7Yys4AuZiUs6Ѵؗ{r4]D=W7uRrS<͎&#<:9dv_H׻_z+Οx)`# $ y$ૡoڶ[ ^o3th b\P-yø_3wfb/o'G3MO*EQ7Z.WMU$+L6'.oS5p8*īubvG6KSfCdXFHuVN>W#)e :x%.;b Z҅AdcZUS#=ua^zءÅBxmzv]DAJ%;ux-/*m)F`P $ЉOÞN=NȾe5Ќ'QUb:߲8{V]뻏!i}i|\J;{gN+goAG_f%ݗT `ȅ":HKWAbɨve]9|.uvVVv*@+c E!.]o^ǹF5ߌ}>h3hht<ɽ^nzwdCx؉7ݷ/Ӷe Z/B{z,g?ut26N;=M{ٰRX"u)&?E#!&O|ZIO} q9"mmATmzQp5BxXDĈC#&k] >ݳg#݌9d [Y3ױ(M7/LuF</V>8-6*;o ݇~h1&ʎ.FdR2*VɑWp^|#0S$  f6fZ R:t) 5wPfjf^Ai~5wtc}Yܫ9Ȭ+w%+`=/^Drh%ؗZ~דCJ/19RCLQIeӝ>u:*5JdsCK.7v"7yAˬK,B$# 7)"IM5dԁQ-Mtɡi- j[@!OA)AJN:ZXȹ{]*My}9*Qi/}QoTo*(eބT ̚2(),")R$vv@w&Ix^ MbIp|4*_>*GL^(M,)P6ɌLkuDɹZZ≜65#gK9۽aU0*ޕ-dl(H!FYb/αSVE@G6}0QQ2Oz =ݠ']VZYrКG%+ y`Ck!`j53p>OfW?ɜI/4 CBk0낳>䤂kzQ(K! :9(&c?5Wi no婡 >dY0h|V(ylvELD2$DL߂pev ɧ_lWWBc[xo!$OrLYX2PҐXc iCdh'Xӓ]frOPt\mAf txF|I_-/y {Cfϭܸɍ36O6Y/~u1/r=qrLuS{2kB00dЪK3E gЁrm3 sm4!hW(*z6񫩯[2HD&cd%*J@)YF5%7vlMh7uoײ"ƔԚ*" lLhl6#g^ .{o/!1RERV{B5VH${e)C(o[N=}U$:$SD.ɕxW;XpT7[zmtaq-*t:9ƊNQ,BȂm+"ĶֻsIIjp}. B޸\BR9R([i!Jit,Zkfܭ*ta3θ.l?GՅgUxd_"{wqv>eߛ'qAl5A ;"פJ::6OcAMGX-$QDXtnmהCawb()PUmuhu:;+ )#ηnԊ5vN򚋹+Zw,['Uq|I 1`#֣LV?݆NFj2hrk2 GP1b@vY`[o97#n};(SU=5nŠx +a!lD!Vp \V'*Xg)x6D[>x)u}c2d(INRL HdD!jFD'5o5b3rֈ_ub8µKf\r_E9A/'oDD& kY:>KJp`y\ jF^ԆOcŝiY6*цڣٜ5rꝝG#qFѧ8n7d?>S ޷ǿ~tk D/|G!.FKZ4>!Lz}O`B^x:w OI )^¤4~=V9g8ds53#&ٶ)V +4)TSĀ hU|-k'M x(0g4kajZlmC5"ꛌ֞Q@7^jī//'liոYFMt2!T>/F62I TC2{{0zOWrOMwG5*/S=N8}qߌ4._µmnxlMSa3 Z5jy?z{ }^}I̧I%C~?6+1,}΃騘Unɪ uzTd5eXe^O1&B 4Sg2Sr`./.>(믷`2^Vgߌxo:]/^f7^{6jϊ`6q#ҧMGBr%[fsJ˝S.Z<Ӥ%5CMICQ%3CAwā tG7嬭 6f'N^N9duM}eSܟHHߘ&(HSKFB8-,bzuh\۶>kZZAu-{<ԶC2ܕD"#j&|k/ww/8^aY424oMnsu繌ov=!)bտ#JCS\h׮>}NS_}RCD+h9rO q.p@8[O_:>6Sgk6G<Đ^5;y~gæ0|&\}m0ˠ^j* 5`6/~eЍݳ_H~8[/fkMZPKƘ/ ܙVX`x3d yΉx2qj!Sr+$ҡ7:׼%~֦V}>9cx>}K8. Amg'FFv heqUIc7j3&7xVPFM*;%zOxov2aS?v{OZrP_}_qh9}ZTRޠDF&s_:gu)Y= ߂Y[fgÕ}x1B ^ȣ"&A% mm8~v_5YM9\576~Z] V;M/Xw`~gM+ rY[dAB~滙]84>LDyy6h̕}|D km5dTWc;&()9ե{On.KWhۊh**Y#IJTj/狛]R=dN({5AoCmG QMypud\ 2{y5W!Iv{OQmT綕6²;p>/1GF9a(4Z '$ !(BHGXjzL@ xNzc)&2$:as:4h^1[󀳱t f$#t;a>]|ܱXk}SvqKNYA 4K5I0 k4UD/( "DB a%ɄFWI(%qƙCiQHQh. `H8J&H)IyrAN<2-D8䤐 m-7$).g}3/ߒPqwY.jA NC%q1MJ&CcIPqu zU F501s:S`8WV RBb1)sN8-&ʤ E(Xtl}uupϺ]V ޵PJdV>86k⌄ʠ=_yPnţ:Y|NC97,RZX"TqOL3:ϩSQQ HA.v,e1.vk*JbgH*2`YND0»NW%Nd(ʀ.2\St2JѧŜ"]ceDIt%" ŗΧdZtutP>yIICW./ƺhUcWS+8%0r*åXW-|6[F)tOW'HWF@UR-6 f(Mu2J;]!!9D#;m j m*8`kڛ6z(]j -@sBL1tIUF{i[T+ yR  zNB_^ڮZj il.3\JVӮ4Պ 4FhR]!`U1tJY ]eNWRtut  +,9CWn9tuBKz:F*Õ2ZͺNWN1_7}t5c]eTBW|*+MAt./hUADz:A2$T]!`Eʡ  fRշCWsBt ǦvpՑVh ǡv(i]AOWv=<+BY ]eǶڡtQJ A/0r W@)th)U]Rtut!b@P!B#`jg6z A -D.-m+ö):wŝ5w/xb \qG[[۹ZvdY1m{Yl9H}/w㌔r+pݧj:~eNb !B9@|IGU L%<>e$$G@Q[L,& bⷈt= QrDN0 )()~)Uh)tJu(tJP)+:vp(2cmR޺:I2fOn{?F2ͳX0vq.zg}z=n?OU?ߥ7IX 6*2*xer0(s>M8s|#Ho]a.aNynhb(,*qWt! Rm_V f]\eBF'LZQI 8{^0m<GɋbM(8Մ(.~w6 Ulg{4%z[}=U-/B9]=S>Ъ)bjE96oBm>HY Ϳ+Q?Pt/*鼪xP8ǽIQ?Y* YO}1SSexl}6 o|M2%2nAbs4L˴Wu8$>P{{9x13{ #XnPg(S?{~}T[M1~6Ë #r0ȹ񎁳ĕ!EC)y$[9!i؀G~72C>Xx}1W5?'*EKKzL')76H Sq>I!FRР~fG"˪}\nɆ/hVW"M (ͥV(a_]N&h\ho;Uqz:^ޖ+46|jf$#5 e} ٖzE3Q|)צtEYA -K5I0 k4UD/( "DB a젺BpdBSRWk9T`1& }(AH+s0AzM鎦 <` @3iM<$!&'^hkq$)HѠv9446[5PitNjA d=:Tbc /Tdm24De*PPW`Ty[:fN\t B*!C RH,&r %c^8X4hez됎͢^YW-o[Vtxj zRr7zNS%O7NMN/j޺o~l8@6#/';^ESY A"v<ߣN Jӟ(bpZ˾2RE j  Q_st|\<2-P Ggzȍ_yhdoq6d`w91 ^Yf#2~=Y X9d],~_X<|̞49?z@_-B׮lэî٠sorL3?L#?CG #{EtE`z銸־m:ߦ#)9L݋kO/ {)K}c2 &mdw]itABcJ{q3wp?Cyw`'YУv.ݳM#8T҃9g%*  XOԑNpXUEQY$bVarTu,Xm(A(s쐻 *0MwD$IfE)v7+$jc|e86Z猻!aChmZዼ1{D+#yGh.Q41`f4*4W`0[I*32yd[%~wLo [2Qlفo{w cPFU7_ת&&} 4.ZF4BHZXz4e__f(T/{әMY.Ft3 :若 D& <%w:%YF۔G>$,mRVg9r@+텪s&6n6 06~8vnU:9iom W.,]4Eؽ{].=U/Vë\κ:iԶBl.]B.~0 AC.ocEJ7T)KﮞstUf7 * lzx^7W،9浒a:otfMybd(̿E֚;Q^Xˢ'lƳ.jJ{B)֡Pus?9u9Z|aᾼo? @dp$'CڹƢ6Jzg=w{ݏ@=lJ*SYkD98:`}N!ed2U@!*` 2ce J8HD3ef+sjݧޅr:|}z9~\q]g*v4Ba|ˬDxE۝NY!Z9'YB%e36RmXL[@X'RAX(2{Uru%@ggOgq6B&U>JYK3x#qy[٫_=s7s9+zf2R3~Y'4<$y/ݺug*44ݢPu)cӾ]7|-3bX']a#y`裱&FAšQ11`FE+=1 Gl+ۆp,} y)"w"8ݷyZM0d;`4)?>`F,") M9u~۫?' ){m=6c&C7Of:֙7`YSDO@j _ԞeVΥT/kՆT+/i< g5hBLܶaD&kBF& \X?)'DtdRL3ONr5]{tBXK|۷=pQ /FEZ}:A>>j׻h4}e.ŶGZ/3( ^M~"ג_HG`R>0}zrVĶq#2ih,+0¨@ct ,  Bm wUqnnYc=6ec]k_ǽ7y=|i{TfW8į6sD Yaѐ1r(t9g=|"doo=w=ks>YBvx ٱy L]3R龖}_gmnzP9sr V(1[ \0 t^3}DFZkfK5RƻmOa^YK2Go3f~, (٧t^CF2L2D V 0O =no{RtGMZL?LP9s$ (-rV9,:Fܠ SYDfdU7S>4Bj]J\5 76w IHiyyE&΁aeMȻmE!I%ҶLJ]7W{Gn?Zn_{EΪCg:3ƬI>*b'x*]+3w/bj|M&zg^Ƶ.UfP1MMj-!Ԇht`8:Ƣ h5՛@Qb: v>v E>Owv:zk*mtA2Eb@LApDm9adC-ڄ9Ą1aWQzzm鐉*!{@Lrc ESHs9,zbLuDZ+2rku`lZ:N lerXJQ!`[BM=jnCqx>6_tt!l4ggg)!£:SpJ՞nu%мh7 /ypƑ4XqJOg p`vp'k /pCD8d'V4 M~t}^mϖŒCҪ.0K*/u`ڤr"+q50[d({W4(gE|}bG#N~H׃KDzƳ$h H)\%r*EM.5u\y&k|U&PӢ`^违]Yf)0pG{3kΑx2jyߐw"%>Od%[֌̺mf]Y~Ix'2eGˁno^ g;:GUvV;}Imq) KRNT}զ$8'|0{3׽J5g;*UW6d}/?|_~ݻ/?qa޽w?8PIlA?Z%^{W4j5M-4mkz1 f{YۑaX-x;wa&e9xUbq_!i_sc AXTI;RIؙ#ޖ veGn%]"/;P[mFho^ԭ# h`3}Vh/H:O2E. B E {Άk+ Dfo|$Qk1ȬfFfBDJCO7N=6ͮ.[ePM'n^{/g,.s ;.<`U2a;(ԯ(Ad,>CJ'p'Ì3T4.ΕÍ"Cی겭n:s k| 2Ĵ ~.(x~X'#AY"p b'cphsL~ x U:%jX\|I G^&#,F3/^.{gYBUf9Ѓdg1En?|z='G &cNΨTCa ց &Ô襢*gθn܄L[Qv乢mH2%/DDc jA^kvW_2WK5vSEñ^7ary},njz@7TvS=xe<^wI[Aȴy}N:rrݭrlOMqLڥjolr"U> ֠LӹdknupC8z'|d6lc%aWBZFet1{JYeLL0Rl.E3: ]&#U2VgeUjP[h+B'+JfWniKɹn̓?/иhi4lg߸f]hu&Bd0 "ZK5MT#@SQ㹕ŖrDgS!{. G^(f ¨:cfQNSS)Mpod(Xjq,VVG4^Ak@)cR,J?BV-DU{҄H !Gb&[ J$a"R g,# Diº8aGKtS1b-P5,""q'6m$鿂ҧΰEU;;I^u|ub$ I+߯(EBh!04tN s{8Nr`YK3#cƧȼrY#t[C QZof)Z8`DI&i*;~DrU{YY+W:;%"逋.OFB7(#Q*"HVsaSi !! ؛o?]͎!–Qaw5j܈~\q !XvJ^0 1̓5`+p530WSYW7PяӑoqRMyg'#9\o}toR?>4(%#3rGO2q˒OF{#(%:c-Hl5~e%(}Mp@c>duٍBCّ׍nw'$n9.ULWNEz0) z8iH$EC0p*DfuA:Mm&EC{od!$}:0/qcVwCLbzn?EECzT}~lWWYGo(ъ ZMDT7QCr Y)/::VR,Ұ!kdRej^켂|!cHyl6A5Y Svu؍ZZW7̃5d\ŵz XX/U9{Txe:-nhS?mUT֢Nq44{Q\FmXэ#^Tq1W57+૪6T"'?)gA>ve7jXBc4fNʶ}uk*xиnpA/ 䭞rEc]*{$qYy=2B}*G `1$wwEHu]vsPpCIǏ,BvVZ;4z{fMG Y:oeg&nu̳41=tLO5p}êϝn7)Vd}ac}z䚥4hq7>>_UUI0 8˽utcH %8l6_6~{6}t ~s2Ҷവ[U>3b44KƄwFdyLko > S٫2K͍mٿtOk {U{p|QyC3[#%<ϸrQuzyXJk26YR 0KʀF4?Nqm| ]]ڣ*uU RN ƐsN{0ԲS9Lx p:XGy`dlA4Q 6P3.> dN d$EsD=]I8{}r>>kLQ.uz>bl X'zSq{l.Yz݀kF`>j)xq٩R RBfCVH)y ͕"&[Z*b)79%u6!8!+{g2ќGN ^ سm6eF4"C)1"zdJ/V{MF3pˍ1P MI#sJMZgrݒuPFk!h Q90UƯ'IfIX ^89TWAjn9e2 yK :yʥ04iH-7iNrcLKsf]wwⲣi}XA<(Qaj;e5&As;22։zvj(=Lo8!:脌1Yj?j 8hPK}AN̑8֍QPȉ=S,e&im W!, &Q' Yglg^}Ԭ_3Q( ti*9 l&`R58eHhs,y 3AToe@N?|$t+uV0"\&CSJV¸^v@ O;Ƙ(A,#(m212naEJH p*9e5ݦQQ |56udF1k .Gt!I}ʼnmnG$!#Y*{0Y_)َ`@=IJex"O"q ZhK 8{7(S0=VECK⒆u-v:\xlW^a& WQp0i9{zB  py\R \J \PJzpŨjZ`WT@T.L[9''yaLDI~Dx+~bZFT_#KJ װ#i8Pm8WWWoR#A`Q\cAZxaji, Bq)BiyJR^ \ EluW >9(W(}+܋^ \I$Uwp/ \ \@JMg%"(rO'𵳊!*tNV 񛫸޳IQ|M1KgLRXJ*ïxxl!{ ;a|_zu7SY? ,>@+h*޽9 pC٭$8 \X4o[s!~9LΖ@^8Ehq{u0'F0޹Nv?w?B`(*TWEZJ}cZk|oɶR<؛K-3#}SN(_U-f0kͫbv2:;nQ2:QJyg9ރh~\1p9Ak? FfE7o*R;MD_6z W1m?8+[]qo51q>'(̚"e8NXd0C4we#@h,RpK(A:dprªbޞoKbʗ~> z:X̥4gx>tS|m3, 螻ȬI;Ǎ4ᣁu/h>i୍#X٨h:&5m9.m&ג@(w,yn_ӊ&秱|*0qnT|9Nb0 C1U"07^:>O[RNP*ߘg+k7Fk mN4-J얦uӎK-Rj% eZFXF%76YJ u&&0$+9]T^Gˋ}TRpeJd[,=gT0 ?%+`JbJoκ͍v"<ɺ5x#(JdJƒnWİ6+CBYA 4&N^G8[WB|asSWt5&X `$0:^!ϯ(j%La孭oAw~*]iՓx`^?{ӹþgof|RO4.&%"(J/E4ӼppO۱8`k刌6_^hI3]fLL'55gS/^=&s,ťtTS)Sa\N񛦛YlCyS_86ֆp;۰L[I)]8ƮNL|2=qFf$ viә7}ifI[v,;__PEצl9/:LJ +dOUmșEcUvFmy8=;I}pz6ӞOvfo[0/?.Yωl&<_0IZx2fR*2 hUAAr KMD ESesW fҶ"Bu!: Y!RME؄Zmc,;"TU.ȹ_oYo@^7uWX]{5ZwY Sr]1Z/cYzp&l-eRNNϛ͆x$R<ڏmD&7`Ґbū5 JH5{Tz>гoY9nU6dH'jygoU*`u^Y182d cX&_=QE*V'b!B\MJ^7޵6Rws?ຮv k)9a&ys;Nf| 5=_s&8B>aG@o#,3-?aC 2Z`QEc &&1Kۨ鰑^ouSz ubq-`9 Ht.(!#FeEdVmBHJ; DCc'yqo2e+f=\LeOv}r}W*M\.zvԓzy:;2HiO=p\p@TbDm52z1FbWc$*4r6B!FS4fC` =\л h;"gCK'|n/Cg8d^ ϛc/1W>|7&J:gTsCT5#!AD@1dIk5-ͭy.c&J{bp}rحw v`%/mdYl7bg )}-(s+d,k+XSbˉXv||&y}soV]v4a^Ȅ +*X&z5˄ܠk)X֨vy,kTb  yv mmÂoaѴ;~aH`mz#4ף۪3w {~}gq<e7ˎ%55MNL$1xw[-ő6j*خq%^R1*zY81Cb,yA)«TqycMw0N6FɄR2̦<Itwirml:G>$˸*EzԊPBƴuDH@I)]+퐗lEᾤيe5^>k2 Te6^Bexd_}B^$BHM8M~hT狞$osgi+oϛkwΏO?"XVoht |ͩ}ٰYB}_`ԢGqʜ.[N׹_+/|sq&1t_s?Mhrq@_v97=~q>?${/Zܣ2 #E[GB{HǛ1Ygì3"LOIVbOG =2xr~d!7jݣ.'iMsY;,Ki[a[_Ikj@뽓!PuxdǓӏŽuo>7AwȎ'Y-k# $In Ю2rh/9 A;qW[\~7Ɍ[sOO[J 5"44"Njl/1oר/ yB@Vzd3>TyMX߶3 Y#;{boV|61ZB Fe=ꪘ!*_:R*eBW\A M`G鹝 K7̥r83DvuhNT.f(H@EhVqʓ%0ea'J~彛㈎^^=4`4Dd-:ہ9p3unY7pjUMXx9H!C` Z"g<<=JN"jVq1nx2 i=$OJA/\gR% IFSO<$ڐUd8V@ y* pNA'8<8U NAERie_DfE(Č\sCLL39VҰ>,ź-Z cDyw #ԅ ~ᜐӱg:xd4DVOQIJ AxQ{A: w{0@V*j!,XNAC4`O.)Iîit{4~\ݯ'_ E7G?+ſ|ߍC;M7-׫dkotɄI5~)=TGsMґ$(A>'+̈́>ŢJWLG^&ZHfJs:4Wv6ofPE})~DURU7e.krQv ̧X_Bsz ͟{೷,+8"fcf]yU: סra!W#|r;q|ĿÜWpqT%(_I佌7I(>iCHK쮚si,ft6 t]aWZ;/ba U/{n6pe/F ~\(Ґ MnL9tgd;Êfq1^CA)'{S}9ܦq]*-o`R$EzsZQrn TϔB2)E,6Y;bR{^hH)2Z+:8Y,j]whJw^փҜz&Q krԒ1&GiGb=P& cкк`*к~=T Zσ~U. L h],}jO16pV|??'m¡K?'_"XWwv#W'w?0j<.M6è|c2V*K\KyU9 <2;;óW5 UV.w" @m%C+~t8?;7bX^5xxo n%E΋旪Vbλ*e@3QbߙNFӹ=~N*_"3wk/N߼.v2ryRp8`9gsznAC)WʗS焷Bs ]>NHuRzYG'݋vsY=fp&r`SgI(C RRCFyLOWITƃpe0JRz-$jl v&wF!od<1T3fJ<4T!᭹x&*%y]zWey W XV -kpK3h:hun"9hJQ#e.ΚpV}QPD%eh tFΖiCagP* AGyB 5#׀C[ӆyŦ 9&3;J3^F.=؍Zşb7* N=ՒSWH0#2dUV^\!A\@qi3HzZ%BXwŪY\Fw+hB"*l&9L߿W^:KQp4>()+9-(9' 1k|jh|{~P􂸝]]ϓU^L(y; AV3y*Qwx `?;3+(̼JJ51)&ͳOGK6~ >7WKW VI> ZW U{ 4+Vj9 WQ~;)IDu̕YRl b:cql]62"|~J N"|t`lYp wL,RpJDfaA\'&Icq.fR,-#ҁ38K3>h U,2tR•K^ *YiRrPL H0pv2LXB3ݴTJ2XB_i +$Xӱfr9q*SٷA\=F'8N\e8qrOE\ejyU ^cԓ*`NG y**SXURA\@q%)+EvV6% nJ7+reΛqiK~f/^ clirpso׌y^csgtV0HseqZǟ"XƇZTZmץJ- 4. Nِ[I+?gݧecE!_Q1򎙋'X:x]o WR߽F6ya41RSe:#q])'ӞnZuޗ=>3ٛLP5'`4MҀ)}:y[S7]:3R(0:z> HRx3Ak:% "!T~d<ԍ>}=i`|\:¶p-dXa[jmPy¶N*7! ='L.LHVɾ;ag* _"Ch:V_L/ͨ a+5R)EA$M21P{l>a}(޿PGt/jGD"rmJ'qG\QDt{ϕ;)O.$t((- 68N˼>yC2G2#[/ A#~&QF?* >[@]مRx9q֗x~zcaS;7b9$oSSBe`2Qf$Ey: DFmCہ ہI5/mZt*n/Φ׮8R s <57`"b@eݾ >~X8;.uCvPaIɥE*Lyrg9?9Th,IU)⥳J "A9&;*=&ΙڏZT۽ ) 8 R$b#Z 4($gZȍd7ǻ/YsȗY$G3yx=bwh<4R7fWE֣]6 >M٭vFIص\/hѹS̄I8a1g@ϹR ]L"dS95q* ^ z]rپ^xe#rCCDFi,%eN*t֔z `HzAwWyl O4N/D_va݋u\UDEz[L1z=@/9RE']c1#g%g^OW} pZe(:28I` df1&~X6l{,Yz] Ees!U:QhBKy%P+u[=ziCٚ)+5_xuv1}sR%ͫ6V,%aTN☦VB~rKx\9GzSVMz$8bX(L (@|W58O0ʞϬ8#d Ɓ`7DKY]ώ {OY* |bNen'ȨX/hކxf(;-&~'>+c}Dk{k=JPR(njEC.hHMѦN|#MO/aTNWwՕgj[WMF[T6w7^- ׁMbe֬9hC3]H5Z]~KҩF!7$#mofq,oIhcOjD3Xfb9ѳ1g8*gG]>Q7=+w撓feS=-e }LԾY&*j4FؖN54 @߾|?ޞWf`um$<_D@/кi}v]>u;{+xaܛWKn-@B~Z8~;._ģ\d *z Mb~*m5U_J1 gV!+ٌꃗ0m&φ:([gV*_@/F\g夳E33PtBiXN}f(.To#=fR j6V p癎S+'{Jl.KeUV pGFI0B@lə,dߍ2wg'3w`cO]ggCNv>v8\{s v>ێv|-+=*MX̿c~1^fYDd\;휜Tȇ3qחbzy uIJHͮ 4{W}jDsפ4H \'iCY7u>g}>{uRe9mS?o)}|tv|ZJҟ|`r+TNʻ氦r9kmU2oQ ucXnz-@j N1Lr8+QEueuMkǛ^< 7~Dr>mmŞ-ixJf~NRf=umuw-fZݽSlGW<ޓ.|:jm# Kw ԶޫG;:6q?m{,+ìa;(f_lhi]ܙwRY3 ]b4ݥignr OӨ?*VE.s:I,N4FVcy縴Q ?*f|vXAwPGGd9Zq g2g+ Y$c[`1:!JW8Aw3^l"! 4:ą*l }s?LQVr:޶A@шڢ@*!hK~|l@dWRD_Q+eIK Բs%2"8<F칋%[H"(t69!pw!ʒ.; K,E-r3rq!~xpD\v{M' %Qڶ-'/|KwZW|^0UW~=^,xKTY|cȏHRC8 +R"2&7p @`sJҘl{ 8@V8UdYu@VENK@:AX | 29Hխ;pbsMhRO\]Q'7CJek `-3%qPZDa`,0j)qtՠ6rIy9t*_ؽU-lW_ٮiֱ \{1ݍ;}E勦.tit卟]d!])M"/%b1M#W'_~aZ.>Jֲ&9EMf(+T!!- krV&'IOFt`74cH֎t6 m^7m_I^M˟to+Zkj:~n.~Y;uTM'_FrZޭb:5^C,jI핸 K%NqbyQM{˙&mȠKZ@}M[m!L嫨i/!JoBD,bHQX $X hf!ȵ6s3H %41'! uqÓt@AzmVŽtEnsC¢eۚ^z@*3\^/o,w֥ U.d%Q%Q5BxH.eEBfH8սv&vg\I jN1&4#19+J֩y0\p#' Yrm21yVV2mJj(VYg)gW?F-Y0Q:2ˠy/%g6򐕱2ʊHO<E D/]COn'&MJh֓8D'%HWd!4c\H;dIQCΠGILj'Ę}! JR#HqFfc.hYe4+NAE}vϤfԐ;"dt+f\YiY45.AyL䀈31CK[w;"$KB/zv(W\·,px$D!"cLm~&2 dMK)q@vN'ՋGwc_tt!κξMŬ,GmvQAdTlD[:$q7N l&g>n>>6%_3|ց'J 3@2%:wČ I>,.e%@D0ro /v:  x.])/"$ D%vvMJ4V[f`"_DJL3љ "G9x'SAÄng̵UeRqn#/o_Gy2x"s9eNLlhbښ9fE,IX]ѧhsޤA Ϥ֌[3vU:Ӆq}ua;TXb/ 2pܬwo5.f86O~`Go'zb!kɠLƋdeyFDiHPl5NqS-dΦb .xQB2h wdu $RƔ۱;Lq<wvVIQ6\#v"ALE pbt;ՇYiC*d!2,97\"oٶYd2 f<ާ ~ZeIW83IF󡖦k4awUUEH5rp\1WUZ)lùSEP7ؓ }3Nx,{Ⱥ:15PA,?HZX.Tb5#IӳŢdB])81QOjd&,/8I‹׷ )ٛˋg^'^|2_U!Q̢ˈ ۶*FLqڊ6'^| ^O_1>s P؆Yȭv ~|ć/-Q̗8nnZxՏV?~smBշk w1QRXq1b ari5E{%wzل\<\<%Zs(dQLѲ6e Ķ3BEsIlBU-IUA%EL Ć8"Cʠj.JsFY Bφpop_歾8;M_쀖+r[q)DclP" \ {Mrdű1 +D9/!X-R@k ( 11\e1fbUi Ζk|7A&M8SŐ\T1AUWR>*A-,۠)blI\X_gJpsqu+TT!P:c |KKS .9TBdQro Fw< Kuld(D5[KR)mΘdCVtKJ<؀t쪤6a>Tֵ>*d6uV >  % oCH _}b%vA.0Pa|~@M?]T\th'`!X kI F=KR2B^#"Ӯ jΎμLm*{9nxrM߁|EP&vw!S?7COx>at'oE/y~ᡫ23>ٚJNJ.O5wO+c¶xOϻiVCxq~\=MpTWy[ۢ:kNI23~W 6=OlFsgKj,x ZgCq:d PbH'KSk5Kן>MVŃJC cؑw6*2N)8jt-`M*WIƂVScŨD-*Mdl7ׅFMvO mR K&Ԯ3.>+*9-CW1N?F^x{yuy j+-slvkQdQ^5qr!XUW6O"w~.a/Y!`H c(PM,Z>%5Sb愩zF\tlZsA:qXKѿͿ]R̐<'XflyI9H6>~e+lni9ɽߎ~7=p~|Mn{(gw6yј"_ܖ@kw`eq5#ܐj(LT=T:WW#qRh;EC@iOɔz)u/WϮYl4>zK`ΖbW&ah8VJ<U|EW ƥJUVmfhgUabB=Gz9%xvN^?YuMeX7 .}] EkjǷb(A#wуJ+m3QL޵,(zA&/\s"f&2{Xr7J7;w|5'ĊS5m2_7_~yvuq(5/w@n;x]ɸyjGعtv,,-h饇tCyb#diԖ]vE=jh:]5':BaDt%a7c+A+:tj('(ʉeԈL \F35tj(d]#]A~c?"chug%i1ҕ7^jDt%_c/jph5+#ȼ_s.R U"]~Z>˗U8o.DXVnًߪyVE0Ztop5o[~_5߼NWxLYW^opK\{,|Q m^ݾ$GFSr@m$TE/HP>0G=eSpqy4?H3oщߎ9\?>m]+AKoT'ۅ> ںwKOm}-B\dtF8HV|=7/]YЈ|إ}?U'_|ZQWy.t b.3_# n$9X3y0h!(k,k-{>GbCDŨ̑@B2FdlV[ 1fH XIeҾ'}~]cF*iA=VRH,X\3p&,A"P%'s@i-=>R)P핏,dbN T#3U05C1 lZҴ0UǏŤZ'KZrQ @f]m`MKB0*褈ql!K dgEDƬ6911Zs&U<CP2v6H40#޽J^Se$m8j f ڰ $AmDtP:hJW (m9c*ѢUք\R@rRggʗ{LD 6zgY1IՇrsyRyXeɇ:Y#5 Ys)’u>F䊑 QB 9ޞ7BU HXR( LPdmU {U 1hɓ}M.o5Һ]a Mqe2ikEMIuh%ĶO$&C H!jAEB8a>cE[&.f# ʸ80ɤd(̨&4$: 8 4 ( er HM$@ +(ӉPu˒M%$QcAOuH+j+w\ 7wVrE|kB X At0(3m[(2#χP9+:j,,∛wD )_J&:&|K@Z<AN:2]:∬ @M 4ՙ@HqHqƂf^ߑ@$6( ^ieW.#{/hZëQIQ@IAPDVKt^`=/z dAQh8PZMil,@E@H xH*мGwXZ I@<38oGܠֺZKQEn"#Ƙ͋BIJ!ĄmB0|f*8[o@ÚCz3-g9<]-^ mzZIh|TxK!#. myT_V 5ȤuՅ*vdl.]g,XN5~d}%y E*(8m`^֊N"`ǻ?]-y. O&!z*}4 \{#BK|t)}юAjwRnD%| \:$gmDT2k{(%'6ܖS5Y$a @AK^ Bk3h NmBh[viXp\ @(Р pd..h]qcm\g 0S(ͮJPƁgUkUkPaRB]$IY6 -(jS,КϽ/tXI#,Iƒ5$BYIe6Vӥ*^`^Qc-A6@*tA6 |+ަ%La&lmRj6NjNae+M[̀|beR4/$0#"Q ljrRPkz&`IPrm⊖4 V*^P@\0 Ǎ6&TS.*w. LD/ caVMZKH2bN KOPq$,M;yۈ'34!wrt `j?zBo}u7VRZ:Q.A6ʐBha;њ?@NǣG0};!5WM19ݔϟ^`3#f.wbn*d_]bA+߶jz=~@? ږ6@/zcΟq2nI67/~˓puZ5=iX?v:|2{gE<9]ݍ_B?f{ZQmYv UK ;v{20('JTp))4"|:J Ч}VGZKT! IJ V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%ЗrX@>)Q!#GGV RyJ V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b% ^@08ꧣ +(`^ . J/G >X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b@.''"sQ cWʲKPe.(}?0Jneva%+4k=}syWBSf&'B]jv> S;v V|MŷoR (V@Pb+X (V@Pb+X (V@Pb+X (V@Pb+X (V@Pb+X (V@Pbԗpi}Y=֫_f~\RS*vy};^\_P V׻u:[n ~Ѵr) 4|:-Փliz- nX% ඳ|UhmiY߬(-lYv)fk,3gѯ] Xl^/ίW(_i85͢?Sʅ6]L痆/ fcz^ 췳졍~CbD 5!,{[ |p^vRJs( ;c>[,f7ow=4ZMI@_?>F?˵Ůb6gRNߎo7a}hw18ķ]QgQw~>Ѣ#M}{uwӂ@!/o{@,zLR!H&J2*&2ĺ2ۣ۾=Bhd{ڧaR5.l׋*,ss[+g&:R RR15.É;b)4]Wb6=Dux? ]#{(B|"}D^࣓}y[A"][a |9;>j=>ΰF}r|Ϟ'JE@ p N }('Aی dg5c=iRsZіV\UWcRMXMmӽKk&1׽04Lzg0DijP4.}~0M`-w<׈ttlxTA=Oh.lܓYK'sY]2j^.F&guILh$LgcDRjRSx**:O"\7> n8C2DSV&1!%Gc"ӴpY"ǠKvFǠOL2?=\Џi"n(ǧܿz&_K*̞YnOtTW>ui(59'WG轱 J{͋H߻uQ 3bNi?Ǎ4w_q{6)^J>>Thw}aߧ^/7o4^(4oO0m„`ޗF}䖽Y__M5qIN'9(SrNUVH%hhSX9d)H{ӥ h^Z~e @& 0VyEP5)d0͖jJT2>[w"&K D=MFO?^W\"O>hB#r Ispm7c"14]RtN)&֦[nNn]YhOc>!(u(8˶lG>pg=7rkbCU1RSTtjmfz^TU {_nґvrY$֚N .rBt€L)'#>!PY:iLCp]ƨt e,!VmJtaQH)mbpŵEeiщܗ ګ Wɩ&6QM*Nf][o#7+zav699@]d`N ,cؒגgSl],_ZengzV7,UŪi>uog) %&NK)i[Pxªw%tnϟ>N$қF8Ni5Аc tOp9>`hZHx>(1rPƨL@ 0 S~6Kg`6'xvN]m5K^r |*G*t?@NŚxE669t\8(M-|41%dܷJmѾVkлwKPR(nԊ\=9..i8#]_Qel\э1-[Kh>!_w~lo,/|4-fM̙m<ɧO'zn7#t弍: [Oi|G:''鬫&fY>4fbYŧӫDϮ[G+g{] ׮zg9iVi%# (ӋrF<*j`4]DC -xx= Ho?~_~~|x ?O4󄥬d~y?;]ѵ5ZXK׮樗CrGwŇu3; G/{?pJ>LKIbQ6ZWyTlК<$glGTp* \3VX~T6 fk=m/-Hr̆ ^ÈMbtӚxϛjqsre1d bIg)vNg8YLb 5H/lX{pX+ئ'ma= dNdL!Gc+,UY1ef9%F0&FÞN=Y_?};eg7 l(lIYX13Yu񔝝;]x1WI!a4\>? ]6J}ަ'sڔN@*gGP ƳLFUg~erv3^C+ og"HY=`!Q>8+YG\Eu*fǼZo,W2Q[K5Rӭ CЖx`=O6[3~ތmQ6CzT @@kY{oBs DJeTdA*2dB΃@^7.hF!tޅ(K\@f^;YZ&b5rv0zGdYd L7ܸ}$NsZZb"_/̗BBHogVMMde +٫1oYěA,jV5 VYP̀ "kJ򯍫^OLRXf}6~#]4ZKI8㺂-R"j5dl 6$qwݪw~߭wɥPD쥎`blX w<[34esf TɩeZb̜…Mڲ~\[;Z[;Zi@L$ZfDU0% nB'kɭNUBcҤ"rָYFpE$&A2 !wcmj<,kk$̰0ذ)ɠ>F&_hJzݳ7-tDM=Q鳴!J6#o-*d)b.JV3DkDIb]ͻ>#׾I fؼ( 9d21NX $M^0Q=Ѐ/V( G[aAmFչ$\H68Ɠt@AzmV jYthxJI/:o(g{B%ܛv0ˢ=r!,*HHBxH-+vBfHNph $  ,&]e9Cf@`̠Iz=`zj`AH%*ge/ \; 4kS26Ј<++6%cI0De9FΎrֿrW-JR rN2htKəWYV;GbIL# чc( D,iK#s:A<kpz~!{Jguo1z9XMQm,ZCdWJ6qmuU>zqAhŒH&ySsL$fTHb@gzcmQ DP dP d/#{K&~1Ng!SR=Y,Kc%; eIpU]uU.eo0/Cp eYd(\ީfBek9;VGlU},22o')1RYN[sn-’Bb˘9a,e`JbUAM<}^"uXvJX Z :ay:\-rvpK?]bd^9dOTT7_\ N"[W,N<9o;E?xي;L+aUaV*I,#3 0<KZXȦL4&ǘC`VyZl8>E&E Zx%}&%U[3V#aXV qƾr[A>.\ST6,qf[pqv5%h˃lf퉒dHV&Ι`DΑy" j֓%F).TŠLl*Ξ̣/jShZ-.L.r,yʘ2Cc+kjPgvJ>Na;d|jq/ʼ\ +*řȠ,4F$Js $#V)-+GA#\i]8\NI54Or4{\d5`^\ޅVG1'kpn__f-Pϟ3ELgWFOe+r1v z@VK~>Og?EOmeXwSeRh2^d9=dTM'mfڃI|:})g=UkR BK0]M%ex*ĔlvI ]s2˱~,-٬N$>^%Ʊ4D$,a4Uag38?;0Tq1-ﮝ;Ö$^SJ_$NZ7oԕN}I=aung:wO'xzwӜGF˞ MWwwnf5we@ydſt3]O_ܣ0K.-ϳ[\Ex\ޖ$x>{$-3#IX[(Z%?tj6=̛mo+8Yj}eg&iYF蘠eèMMxO,+l&vQcXS۾}ģynHeé/DԈb Q}4˙[ysѰUadeA'`.3[g 2hf0n ǽEF&aN|֯}˖Ti67x˙dpVx!sfJ;k1u`RnLcȃҫ bꔲj/V n] GrJxPJURHUqEa_;$b]Ýc`C|%7w{=k{f"qݯ44"%Jd D]T]OWYjpQ%*.yy?_rW૸ď~?]/%] ~s4)ȏ{?֪Ӫ]T=N%IE0)݈貝]9/QW/_9 ɑIKwZ҇'w_[uT/yZ_wO dÊvGo_,|eֆZU)7we9> zǨ[x*QT hql݋Y1z}Ynןˋh?1 < _qnJYJYʬ"1)NZH49 73aN-2} Ⱥ39 27|[R6돗an=4}v9JL[/aў33s%iHrTiSWd1<vR;>Ϲ,ʢՔL;R/un\h8M2 Ȁe]GLNߺ SKNK0~:rY)E.W`EbISW"l] $bP ^#tN3kP+|ʹ`# TWGK;l}"|3n}zoʃUY1by66#oc$v=eyj&+2񣬐yaSk$щ3R;+( zwyJCw[;gwy[hdдzꅱ8f>Oc?ڦUd矮?-?CӁNf,wŹi1(0uQ^74PD‘gLaPr\DCIKZʠС>Rn N1W1t.ɭU@U.7~,wW"lY~'zpgѡ9IY:"LY1]&^\)Xn\Fv,ˋH^NϺ]%L}.w9;Ep-i#tjFUx߂YϠ/o]wQ,& >WL#˞F>[';_Bf5O$.D.b%Y*N^ rA)$_ee!#Β :F:NK8OJ2 r=/d[%}0tHT3,^ې:H`fؚc _wIr}y_Ex_Y^E_K.9}2}eW\8@ڷEkxTڅjGjUW֎fRd_>v|am*T9 ÙҍTtrq踮R'sOϋ?~K:pB+̎3.Vq" g_8"&WW6{М٧.YљщtS%s-K¦̉+X2+i'H &;<> :TR|Co*ZXl EG`Y(`}j [PiN T Uا~ʋM]yodiUrЎK'b0q/Hg;ozH.~3RR*ߛ:$^Zg@gπ0K^cF0fi@7ȅ2Ɂg>˒"> rldDkȤPBLsF1pr,Q\&Y FO-[pYdcqn4lTYn_[ɫܺuD*Nrm܃pOp*is=^o}}s[st+'UI8H]|v[P=v6; ,ηMֽ~ۧ^7w;׮|>~[?78"kvB>H=7^mwy4;]Z85/mz:gwdQ ۽*:j#il\k 0gFYL0aP'Eps*S9  lC=];r" &\W DiK0BPpEj WRMqFp FJ HCzճkhz7cFkm{UF3ZUgઝZU; W܄CM 8` H5t\Jg'\Wyorf}n*:ջyZV%\nV:k9g23>Z\Vf6?qcq7v. fm% (X I.P0jӤRbZZ@+ Z+{ 2GZ܄⊂jWt؅3g`̄+R{qE*4#VȀp W$W]Z熎+zM!p(dH+(3 HhOqeA8 #u6\YqE*p5B\9c +oFj:H%L1 ~*r^(jӧ+\V* ~{^/foʰ ]Y)k3Nc-%){ ^8Rr^Dz{u]+Qc(6e*+28FƊC$.Kzэ 6_"LYm[a13w>KpmNd!>~\3VLpl{0"Rwi"MnQZj"e6_&`z.<GSIk!ufu}`ھ,R N#\]r +m("W܄1JW 6:<.b9zfJs-װ3&jrk-ͽv_e_a`&?h/U*S\;*Lmi s-^e=qƚߔ =nmp8L]~JWk.on./ףu)x?٢1ZCy9[BqcF;GXְ~ɲΩmcd)&wbwKHtHr{NL-ھbRiF8.J6\\ =uNqE*݄1Jsc W(؝!XW"xWΌ* W#ĕ6W(زpN\B=u qE*ʹ WֲOJ1x6⬙91)\,ye\oڑj,IT%u9k'\bzNAU@R+Y("V W6j[ H#]\)h{HWrKi $WUS0iR9fσi& \`gfC|" &\WJCHu: P|Ry#C;Tx \ Wk, W(~Pn"J+RkqE*p5B\ ! l7?`ε).4-ʭ1Jj{mt(s*o fؔc9MS[F`&vuIOu34 eheedPr-q,:PX8S $WPH93 R+ <\\y(" WRO 7cĕV]`# \\-CujBMqӎ+l'Ru Z·+R)aqe4P,\\wjwE*B|ypeqF+ "}Z+RĄAyW(XpVIPpj R9w΃+H HŪ+R+?$z >\3FVW䪞72SWv*5=<$\`pEj JZl Wg j}wUٹsfJsɘz ap0MrAiTk:I0=FL+΄ K!ɕ.\Z CfqE!`X0"܆+R++R݄  6xW$ J+R)r+T8xWV` WRq1jrB1e ++|"0]W j;W(=Sx\ZX WJ641Z㪕`'N9T;fTmd-p%'\jz.. \`gʕ̅+R+qE*28,BZm1><#g #ۢ2%4 iwvj ӨRp=azJ)5|RJPKU7|Rʥ &Ti=Da֒0NYK9i'sj5MV'kX+#'܁ Q0[gJx8_!.n[?;HAFmdxVzxh(q$5LstWwUMW "Fj#b#Чbc#q?9uN]]e"E)+.#RWH0p4*ٱ+oT2ڪ'$7FNF]er9uJ Ъ'`DBL8!H-WSl7} 2b6}ZB%%,F!Y-q:᧓8nu~?-_?(9?ݕ/* _Wc^,`2 ?AC%eu'z2Ղ)-Gq@7Uk Ar|۝&b׎*?^gq3קmlϼ* I]%?n[u (LjR+oU4+ZmCwj̹bW7Fc';₟W'R!OhIB81L[7&ƓEz: I.fw="Iw=ELz\-lW!(䚡-Uƛ[ e*F8۟#N; mΆÏvx$cA Ɋuͭ,8 M ("a]l3ΦπLrCEe5YsZ=9CUfM88\M#E^Mp4c HRX̙CDN O,-)1ZjX(*罁]XC 󫀩&g+2Yd1R˻OgqЙuP~BkI~7Nr_`ԭK?_ >`La-meb36d7'RTuz'WffPToyL3$qj狵lȃ?IUui+azMB=%C kogk [.a'ߏM8`F&y9A-O216 cJ`qmaMЧvors6}|/RAt),Ӗsm.0BSJHGLͼMmk$XЇӻE2~I}lnRga_:{,R˃Y@'@ Uk{(JE]s60-WPEsy|6b2!UM?DݢL 7 n>ZDś R)Y_ϗp|,{@,Ա }kϩ"hM,(8 rft {bSWS}$9d ^ #Nq09ƥ`,$C(1L3#ymr13epƣo7k#gwp[`3h[fO.5%_}tֿ u`"Ks =UpX, < 71=Î&Dߣ{~@dT45STkS;AgUAdT4%pV2H%b.x@EI$t MR`cJI`\qYmcLD"Fys!Co89[ +mz@m0G+:DP9sD;<] XwE֚ϧn:}fYZ34:&N2GL H!FG-,@O7!$!]h gU^k]ILOv`sэ*) hP"Td"KGLA}d(dqIJyt<[a]֭_hED P- ,py!<լe*r2M0#Ee3ƀFJ!Ɉk5q Y Q8OSkGM14q0˻xmPZn)x)h6Zz"t˜k;vn 94̫oi*ΒsLTrκ8 tkC 9gvC{-É핒{]׺SW4Aܱ@~8._~10u to4ҨD׼`~{3dǿ:z݋?|?_=_w^ߋw~x/8sD- $ (}7Njѵlko5|M^L2k)>vVlϯLn)@<|^&~/_ G9ǧ|\YM{peu=F1?_/(U(bJ۲JK> qjҊ6#Jŭ:9U=qRFSx% ~ (m"&f(D#}`Rq@BqGR[DMB$O$ :%N1#;ޞ>D>7W~(43'g0*ȣm_`pd7߷*DHe:LGhmI{a.e0>dЮƶEZAZ@з)5~-v_ gB>r+Q㴮,{]9v"R.2iJd*hWp'La  mU;kƿj2mƟr\BLAaZE)w&ˈ^(d7_9Vs>HW&'ud,3$Fn[Vo9wϬC)IBKDpM+Me F^f8¸Ĝ' "Țޣ$cb(Sr5D(e+InqlH0}3O#p:N3BWi+B4E? iK}>i04*g$eT4WJ7_RCuT&IhNt1*-XR/Q>,9FӇ,RFDHoKB:eYEJyN 0֊@"pBQo߅QX`(p+ec48sJ3p-j jiZ?ZG"/'"\d/2JOJ;^,EI^A8QU1J#C*Sh ycЉHBDe PyHFI.J[s-Wσq'"|Ȯgee-#(e`] %rP%c ,:1 0(ZhNݪrͥVr-[ytII+% $E$ЂlKh yz_r IgBP/V5mJ`saQ\R^Ry\Ef1ͼTs;7eﳲq^M=Bv:@Hx"/ D!G'u(>xIJUt,sƅ)0&,RҀ!0YSLNJ', 7^Go-+sgU6%Yi^V1jPkiɁki6ߦU[gM_*Ncr[trN-C',ZeU&uYEh+W 1 Ӑm-]e%1`s{$1K.zRƌ^D"䬜Mp#ZL UqjXXmdw$XXx-Qyea=@m"ty;UƝϾ-9btk&2Y'$s["eIQG&稅N1VCl 29 z6(oTd1,JfC,Xٍ~6&PP j#-  < fH jdXt!w~!BD VxĀ@sE21Ò),aML,CR9,QM [C85YD 0QUQlDrHA0Mc]z$Q%:%, 6 >E 5!IbumzΘl㤕䘍2Z%r} b! 鄲YvjF/>'\\4r i(9Ee\#.uu?f.XUL@֠P@9WQko!r2Z%qb4N0S`~jQOa^a=WeΑ{~|r?ʍ nt~gmAseoF1Ŭp:$Qq.z$u:=$߳C|pe;Uϟ3a|wT.d!H}Ԋ`VM ĉ,U7ќjM<:c.c;) k8?3|/h<^,(>@;u#lEk)'yD1H*($ jh 0.e#"< "<#0`4\ @L$#ȳ7XrD pM[dq+Ż' &N5PnwZo9鑈63OP RcB:QŢ|$P`gTJ5O͗q^{CA=(>%泜x`A@֐RFM9UVN`aDZhqN8޾0rGϣ:Iɧ`=|1MZ,36 TNDY fjZ>*}ަ=:/A %ԃHQmR / 5?b>Nޜ&a!s!3)Y#J{%@A*( TNØlN\i愉`?M&Hƭ=HR¥ub}MӤMgٺ2"ܜ Ŵd+LssޓbίӥѿN9yK9do=ق]\A}*<̯*M ܾ^9[.ӧ#~(Sv&m׬svM[ϴ+z(X_R߈MR;2H~[ 4O-"Ln4]y#smA] i?-z4kx>Lg% v},5Z)f98{ rrgjg۵}G}JIB麿ֻs=c0L@Wfظmqg[ ;v'-&ݓJMonj|}4w.$Ϝ BKzMvsle9rz\FE ^@S[Mn\=H~K%YawؒrԞIzB̀Kd֚ge1Ŭ1V,(Α9u,0+X}4΢91% 4ԝ2ZD) ^oWgE痋Rbv~ ϛ"Y7iqYޝȲ-SɺnϞu2GFgFY%10ke2j"^,YPM^h,Je)IwPbɤN&&A̔ذyb픷3UͷdԔh IqEK$ASW3S0_弲5]Plz=jytT/D@aߗ|6ӀPīnR R tH> ~h}j4ytnߗG4_=F` /%=<=N\ؿ_'톟H 3 -I/Z $Yy MQճv%5|-dE~#&t$tL4xB}5Mۮ:U'_wˏ̼4qС{L~^j~qO3XɪN]O+=L~OUe& J֋+qoLh4NE>.'\O?#>6SY搷m>ϿWRJlhTrRp&,p;h=FA+OZQ[ߛwYF{ g)i&*`}NY@iEgA!+cT^ͦw-C[\1Muofg1RHuPLMm_6'֪{\8%~aFn;a+ #KY5o,>mvmX^ %lPds)62{d搲b:JV݅^FcRQR|L.hQ*a2hS"9Cd&0p.3lWgO}]h.߮=WXh_fz6qU2gF/9Im̖U5i%?Ē˖j`1eh@'RAX7;r=x&!qrPOgrrF`#ʁ)j[Y߁= ^;Z-m6ʔ+F,KkGO G_X_D;:B'|e9"Rm"WEZÕ3JTF\"3JpU5q,RȮ^#\ШUZpU{,pe \p r#+XaWE\8xc*o^\Tx7xK$E%JXɬ/#"#MS`+@I)7Lv ԧ0\r߱81;aEk7ʧ8ijlR8m3|D ' f=Ajq1ťbFI y#ӽ'#{Η Tmv&WkkxAi9P o?9i5$?$o0rfL"`, %Em fvfOCVt)q &9RrtNgUt9"E"RHxb 'Avy`NE}1V,#r~*O۸, 𦳶O8@>@}K+g"*ɀt hDVhHL"Ҭi|lv}xYb_qUwH7~_Fꗔz(%t5&|ŞE?r{_n?W„S$6"rڏ ^XeUW;of7\!xqE2`MpjJT *F(iZ0D DW++It2Uw(˴tjJPH*nPCۜDϵ=8Lt JR.j]!'W* 72ڧ,q:]euճЕ&YW7?ȦЕ /ԝ2|tiNxo]!`#tc*ՍI lTM*䤥WHWqQZ G9U{Z ǡ5#JEˡ+y tuﲟ gBؙ jFW-]=vi BϑyQ]e52ZNWR++Ɣ4 EAYAIk)s7S6ɪDtcUfBԝ3JZ~4H 32} ItњU^i$'8Vͱ2\M+Dԝ2JhHWP>Y*6ͱ.getQ૤+ &ŮZ8y#*Sw(lҕ#842`#CWwtq*#Q^#]ihR ?WEW.oL=tQ65ҕB&Ů2`#CWW7Ǻhuq(vYKW'Е:pA 83ǮCkT(,=VAWN=iG>uu\xS*rx^]1f_ C!GG [[;5XSߺ@,+_LصGlDvnwm:j.v Bys>5Ժu$vl@G8R|39xWn#¬ ^yU@P8ނR?g=zZ?ʋYߍG.f+o6s 䎈#7x?ϲ,?y7 m6J|By=A[ #U,z"T3_7v.3uX;pٍQeA ('rAn Oxǩ A(CB)(x4]BXvǍ×+B#I,Buq @sri=O (DqŜi80JRQkf>>NG4HZ1;`F{f8Y(̞Ͽf5k7\Ѡ○cec~i2\ OBv}՛ ՅkG\u)goGO)_xg~IHṆ,GPG8^Ӿn#;-m>]Q@Pc 0m& n)0cfAhGESuJuɅfhR#e g CgՎseLEQ@24CT+yXWv GS(GyBC.89 kyИ: m-G).&|MBB]*F}4^HDHP!d=*:P^".VdX 6@uF<@Bh*zy:S WV R/cR^)pZrM+& ;Lc5ұq8umA]&ӟ-P.h\Е)?k>Di䮦`(<|ѱ- '-ꇇAcYAM 1"F鵷'ø냊€1T""c(\Ь@<(B=OM€(.aGbya@(#%RƂq; TpCm%VEoІ6 m,ٴk)F (!dԖF˹YhiZDHw'ytbYh*Ut4F"qo =@5\"Ӊ"*au:hE$-+V mhCJ,&\P$Fp-\T\.qby>7{B7  <^tV4K H=ũ h&/^ œRϛYJped5*1>@6e6 .=/YJC\Pr'Q[C ϴ8Rs'$͟;$~w&q ~e⸻5nwt PӁ1 73^W@]FZrt5'rDBs"b Θ 6/= I6ǟl0fR֌D0,%uRښTg2t&Ƥt^-npR1 مF婢hy8)-}ZSaá|'qǾ(@q4P5Qh=On\Rh-A&*NPq*H\@D dB.@/Q 3),a㨀ʱv*s`Ό_7-#;dJY9,fm=.lou2 dlΗ‡w=}SH r/_2uJO.Sv_`[ϭYNtf?O[زK{z^5~yz^k_ƣўϒ[GU1Wx^wOu$.Es1ECwS_lyNքio^>L@ל|siPHlxspEcΟh{ȌnG=U"J$V}(K|sSkYX{ qrk*5jHԔ _0\p@jlCmFfs"¥1;!\yH<9"'^%u * 52`)kLHڅhp`2Qf$'~+Cl$u*ƹ,炶i^OWo>߫u4`s$SSE \c*$W57:E[tep&WB3IŹT s&Iij&gYJ^͂e*)^8E!DxH[) Ejt[)y{;4#48`jXJmVV2H%[#()&X Z򪫔dӶԱܟ<1-op9l"\E8!FG-cyz G%IȜcFbrQ7e1&qXr`_4ԂLC<Ӽ[vr#Es)T!5L IN,Y O1D`zAuyھ>}:a(}* =K &1ϝ;} M]ѵ;䉷H%}3^\٧wKkjw~X۳^n˜~o޼]-˱XY.zwZu~ 3bDKx%-L5c&tl:Ϙ.W #Gg>ls`a '[5V;9T_ܥ'Ez>Gzq_6]s^u?>R _jlo~{'; /?ƿ~2Ο05wIg20ࣦOдiw4yKa^ %3g'x8 +2/Rj~ V;}2ku?٘b|+z*QN_:@>(ͫqrGjOl@}ʥ FmѭI?ړG26 Q~l I7$ATjcu+6 Cb5{.#jkj91xV7&*7 MRQv'{:wGƣ|xt?6VŝWa80v;_L iC^9*%rQ*s~ddYKu(DoJTXsVN%ҥRNՁ$HWq9*YHՉy/Ab7=9w?g|QmnߦtP;;7o>k]i_ȼ59t@KM"H%9N8 ZQZp E! ۸dr#]ӽsr~+(^Ayۍ)r3@Hˁ[~2Rٙ W5cVʐ#J')+O+람r5%>\JFC+o]ӎOq/x)os90ӂJgY%c3X}͆dQ\颻DmtҩZrt5<'LQ딸sWrf#m$f DRRm,d2 LJ:XEmہg3Q1mݳzpK7v|q<õU7`>9h7o{ K_iaGwguX}c9C)F v83YSs.56$h>r=‡|w䍣G.O jͣq_`lf/g)~Kxcwr{Ұb.1\G1{nв'dVZSP qտr浫ivrYwֽى+?m!g\i˂=&XX恹pqaq5[OxhE[,ï.if\ݯ}B}9jD$mS`K@Vg+YGbuP|8;=&y5kEv`٬ehMsS*fKhxv*(dd9$:*!b]$x(Xg$LJHNc5XXr#.TL.d,q-S|kL Ch}q#KaW*pACqM%%:Ik9xg;svZ ?Ց% Ք/5p\!(bTTAXj!`NµK/fBɌ، $ۀJc͙>bTgka-|YR-])$NE;"lHs _ Lw(NQ!(tWUǿӣ.+|d!5Sv!dHQOC𻗡".Vk Ϫ5%Xшs" jj"ȵS4Jg3Lm1猇s\j)95Ybɳ*.zkLL5er1Vȕ/ >FBL.9 /Rv}Vc `֑J7b<=V \.%醼a*kCLxe@@SzarNc٥DDabB90O=0ن:8ܬnD9rʃMDcA}! y"8jiXLd5tąĕ*L:4g]V0ڊLg V٤<9~ Ҽ>k Ki))/`hg/4dkjA(-"(4a`ܤ w5T͈Xf  l3hB'd4v[ =͋詆$H4M0r3x9a껠^!,VhrpC$*r]]ƸJ'8"FID) e@7 0FϚ%$beD9*ձc[ty 5@ ϒ]I@Srm>ȬQ"Vծ%p A/Gd⪯6٦k$fJd `?A.j؃X0"b 8\-F`,0n**wUB@Vt"BhͥVxA7 #DFh3M@mJnZ[Qq/GmX@ HX y+h<%07"=h6M;<_at>[^6m1vfWBPj/ICjD3Fr n#c00O4UN4^KŰl^kA#yHY-:`4vfcZf0~ȟN? zVZ4hp0)A/k Đ @R.3{ HY'%\U0)#S : 䐐A hoHY@u7odXz4YJid`+AK14'WW$co/H.zU {PiQ08UNFB,&Lg=gU4~ *ڨdra<ǠIN6~A@MZ߁pYPAZD`2 BJ\5g>Y pM):==4kPf|0}}IsoRD ڊ݉0Z …Y'X):T@-$0E z}!?MH AOr|d8Nгr9nn곱Ĕ *5^-M/2cm&Ť!0Bq:tbIWrlI҈5.K,?u-q!"`Un`j?Ɲ\˳t^`y ڷƙ L_a0q@%19]_7TlEO5D*̇˓zR[X\xI-#j-ޔ_ !-+,sU? )ɋdӥe'LS[S*R(ӭia Ai V>w?ml_QK(+|d1~|uT{ `?lR}iO9YM?H B/DǤGFEO_[dK,冏 5 5( >#~ ]!$]+4tBCWh ]+4tBCWh ]+4tBCWh ]+4tBCWh ]+4tBCWh ]+4tB\]!`S*6tͩ2vʍQl _3*F%Ш@hTJQ 4*F%Ш@hTJQ 4*F%w=n#G-~169`n$1Ʊ45c{߯4uZ#rYd)!tUwSTJR T*J%P*@TJR T*J%P*@TJR T*J%P*@TJR t@=>J ѶN%8n H!M+N*B*WIh}nI?Qh|%Vo`M#nhZ1bԙZ1fEEqcjCңC >FkB$r4c`W|QevA;dM!9G !;69!˸NC"q`NxGwW^*6Gl<}ipy=T ᤸЇR,WJ=5D;25fѯ)ZVQqQd9c4U,7Vơ,\] 0t7NdaNK5U~)=Q|$ȼhgig] ̬ӫm-c*[bvRZkj%$TKXFrkQ2$*!)%a\pN A/V6W+̯KT⤕.V?>NϿlixvlRa2_;puƘ(AȄIh$?J[L>xm 3ZQABYIۛgFAoxܘH_q`.:5 v΢oR~~Nݝ'v{ +R iu@bI&(p҅$uCP[:Y?*xM_/9lf㳝a$j#3,:@XDy 5-(X,v3o^2l4whc%6\`o(M~bO0PVޭQ!E yX"9ÅQ#M?Z!![dnj̘ nfdl#X.̬M“r = uRVh5a~l~*T r=^lb&!+]C@Ɉqj q{T5OdB:.tV& \dy&W:`٨k툱5I@/Ǎ nI\ 1FG_X.VMbVkm8ģXFp°ԉ'"хqEf,+Ri"d *4pML$( GZbّ1`͇ܲmP?K8&$ƾ0bq$#ZnF, 0L>3s''¿4:2f|+W ^5$1}jˌ nF1MW:cyeEZxd$qA,D2,TDQ5BzzZKOv]9(9%iv"S2ORU>fyf~qA+" <xcOYhw.&f8)${ohtnE^tЋzA/:E^tЋzA/:E^tЋzA/:E^tЋzA/:E^tЋzA/:E^tЋ?NbgjfFo~`Zr}PwH)6Q[Xw|zJ p ` bC)\BVp PRjJ%.=zFnR]%n}Tgh.QCX}\|yf!tIn>A<f84ޟq#>xLe@v華s}4tYtI`وB<}l}CI{($|+a_3oBY<궄L~5H+]U/R!cw#l5q&f-? SS,Jjj>OP[e6w>;x-C\ʞN`HZ=ܞ),߿1 \A'}ǛGۿ=7?7pw7#^29t8M'YfuA(vȩ?4,>bfge"˰ijn}wYoOe?VO0bവ[Uf0A2Ԭ ;+X|]H/  UդM_:fGv]/lvNysՀ~(a;Iy*`٫-g(*elnݯTr] D s{[-H zb'*L> I 6!I Z,KA$e^\:[C0 >-/)'7KWm~.bGd]mK/XL 'ӺgևL>(yUiZt͵J RBTGVH)y U"&[Y*b%%'um#{GVܑsdeYGVFJhN#'N 7&gQQ8V32o:pѵt-zJra'AkPD%)^0_+6u\k(NHkl%a$+Je%Wit^]ȍ JjJrĮ]5 f` Թ5U07 E0rF. (E0a+, ]!\eBWvmᵧt(-tute]!`5v$  6#t((tj[xV!5|pf: .W祫О:DW=d+SWO9 á+ʱ9P NWRZqb^\VpW7vkr~Bq:w??0MS#ݴ22os!QN(`Kxa|:ॱtUiR?fw~O?"Q_yx%*[ ^Ygb{JɳYs5LWV_ey7# L׆{y#0W#2 Q!mu?\W*Fwϕ+SQKjX%8RTdL8T^I^YYж&WY _]B? ˛f[u *QQW"[y-d>u2Ϙ}4/PuoiKnp5K ڲpK \ 3CWWU_ Ee"Jh萼+| \ҕ4R5 B+k# %wpJn ]] ])íR2(=BV-=u Qʒst={< +k ]!Z{ QJQX{Pz`S#p ]!Z D%ҕ@0j8tp`ADk{OW BXBWyKRi/P#+λ3x Z}`@_KOWs ]w&mY*!YA :$#`FuUuĘ"^jI–(lfw׿_zuBCf+{ym&1]JR S39 y׸ju`TiiĕBge]`s]XWVq*hquP3f<\\r5$u\J[\ _e+:\\~ټ{@)2ixN֕e||Wu.$|/ WtǬ=‰89 c7k&I5T{m 9HWhf=diFBd+{]T2u\ԬWXjgtUL5+%NmoiJi3l0r)Ө%iTESĴOi1IE&L5Wm;%/2/7DFdxG"k;Y9U-@,gumi ,;Mm_ٜנ=hf2b`ͳb\sbZkRb趋}]l!493`u]#++MUZqs@0c$\\sU$u\JZW+j36< ʕLA:Y|W'+-9Pɧ3r W%+TY|+#32>frWq*9-NWVjś@f l&ߢڧx\J1Y("J`uU3#J}\5S]b-TC W(ؐlpU5 rV%e,LZnS4LmӇP',^cR?OY8Or* jƴXrRϓj=),v0D l(WfԻب}]l!<#\`Ih6BT+T+Ak RK2Vf+H.BړJM[quK+,H>Ax+wQjquеZ[W/LB;g1t.%|Y ~Wfg_?t}ؿikG3B+GdP` jxKI#hK=&^F_"Wy+w ~^hWCt".62gZ\EYH(Q]\Q@&p&2 7 hP?g/~G%ݳq2X%œ;mz.93`ٛ9 ^,S!WϾq3v 㴦"_y䩮|>WV6[m}RUuvPu꫋s7ng&8Q\vh,c|g}^FcvKm_C6zFpU1M5H9x%C~ENlSRsJ`'jF/ɠNj% 4FUe#NGbDUT| _C\+wL0_t?B9&ՀKIyn] %TT`uzF<ȳ_iPKvSL-Gy2O?5~jE6V^.y04koȲdt:ѫlCg #Uasdtu°E/i(JŬVq6l2m~~ӝBzHGw_1\?CHOxڿ9U~IwVfv.0؏wϷV3B\6B«;?ױăEhVXM& ( R]0S2D@׭+eAd"l6KV f`gr5)h6l>~,/uΠRMt fqm"İ)XUjS q%|7hbLz@?ayܠG[CL*jGXC\o5R*M#`m:n|:W㪑ZA?[ ֻ:ަ[oTQ={fsQ -KB@_0"鲤R<a U[k4/VSlQT=^niTז. zvX-hc954ƣPQ[tǝ܇ 'hbu:7D6ysa<4gBn al;['nn bygc 'l7Eۤ:=]e<(L YXrj]Ƀ\0URfsEgx& %nIIR@(KeA}8򯳩u_W{+B9"a"Xe2bVB]YQBa=(ǛIyqN=}wiݖ1[o4h.̑0J[b3Z}E>OP乬>AOtOPeTv>Ur,iley ~kbXp w]Qh-]ID>o=FP׃=3B$4%5<+S[PB>$#eq˕Ptqן `^>3u̍Y Q=B h4"uq 袞F rzVP4t<ő.[4D'TyBb"mf:+m/y%(gmԻU8q!ŶȏxݹpD;lz}_RpɌ?M{~ FMUJU.oWNA/߯ͨ]FwwHZ~}‹avW 3dY%}yrNۚs3Ud/w_SD'm I!o m ^pnS/,҆8T|TV|2]/zsx>96k{c2BWh$Mq 0.,|;Τ 6G; TSrK_6Ͽǿ޿|/|2x R`=.*A?ZϏ0݋4 Kv^$r(o nrM&@t8|3 \~19%- [MB\C',PWG+*UR%CtőA5SSDj- 7FnZHi#1Ji5QPnEKO.%V\"8mΆ%U=! Ugjȑ_a]fkUTr]%l%_LjI>]*=!)KF(HS[kI3Ù&4GX`8Q(ǒC#"}M1E8AItJSBȧ pi`r}ĵuq*\ف&q}&g8H!;;/=y7t7VMYϓeFK7tc?{y. x &k]j2}ŌV&;rz>d2ŇL$ R)882$b qd6dEgg8)C A!O Y@X-+e>YI%T&o"8F)!blsxY%%JћwpEe>W.BIə(ӗO툍Dwr3#mepdt@ }WDS/z Џ&J{QRwKFoC^ho' i=BO͑󼻲 MN=ZASQAk)0Xet`ChXL |^ϭ a6e!y|%>C&BHUtwh- NPԢ'MqҍñCJDɉaZ,CҐ[L)RqHy ՊQ88Q7Ӊx\EEj$=BZ$dv&P6D"U ڭl^ȱߎ"[[bNBS**$iGTXm6y3K0b,LC)Z?"j}O߈/K%G$j*.!aJ6yd#=Zg͏.M0OMU7_ڳUL%/~yE뛓}ŊDx..߹}sevAeYJ$8 OcC^J)ZHm5 Up6`/$(J%T֑Opl'Y rLrc7gǎybk7ӎxm?y ޙ8)s8%%* KA+WR"JAۈwl!32$(TvkRdeHȵy1qPz˹8acGlGzDxY#NN 2=JaXUA1Fc'KRbIFFFM$1G:ͱq6Z/ցJNRLLld$5Nd-h#~UGnk\cZl%E/'x~F#th$eU:&xxbE 錓_| 8V1xd#z'v? .V#6՞MEx@ܔLُN(&GɊ |ʫN{v.jO1NI?fQ=ؽ[uܠ+<}O@l61h@x_owzW}ml}KkĎ " E2>{>C0Ij|, U98XN:#eғix4A3yl"$Ĝmb/J$h(8Ox'E}}Aizr%;eЙ?Nz =]yn;6E}?T=_L7 ?]x?'+F n~hWk:moқ"K_ utC ~.r w\w_Jɹ~hhvxfsͨZ̴3(~s$Rp|e;InȠu[0[]̩esl%T,+"74/|dHw $/Vk}hJê5n<;-(9*w _1 WIrvs- c0̷V@1Էw8M!ONN ̤XͶO#K7ͣow6PIV`>2>bq,2eCy[CMx+IoGi[FU{OtnXgFMpScXpqǃva-;mTU*ͱ4ۑ͜dEVf ኔ`&p/kK!JDŁ{%K>x/5&.i͢j\Aǖ-9;YmyW k+ Pp QH cR 1OI"@LNӫH}7.etX27ݶpGjISUUE; Uwu0m*uYX߮2n\߲WK޾9Pȟˋ lţ /uQ|&O8_&-ǕvW0)ZɟldV(>es7)8o/}fS۲Y>  zOn+L|6|9,حK$0 [o{9m4fٞ=`vggv |9_m~?eY8_f^&rO3rfv\|fm?_=zwIl,ER^tE2 8 ۵6kZ#jZhZ QdC.r\Y>9 _d"RG8$>M>"٘S-ZbCT&4[; :XoɪҚ8 ~yL+Gˮ?"+Fs+^L5ڡt_Κ݈cK=>ƾ4^{an{ǻM P 6l$ ^Ǩ)hT֤a Oqʀ>c@ !%JB'P(l]N&GoѮq#PK΄EtGԖ3=tb r"' 4M: uHD0z! shZ[RަdQZLq432&,1"s%kZ2I]Q Tt^e k YhHff." ʱk*!Bt.:bZY5N)xxEdE`. yCF t9CX{Jyanj t.1ɞEQEkWIj\t@!rwKOV\R>OSdEnҺ-HZ#ZsY@-RTLKZ؄)R"90$t6EАS649ɷY=,:L/|=K>_S&8߽?ozgӯEi7]ϼcu䓳+72}=k8)[< $atatTDP]Ͼ34R(BrvQ O=aRO1싪? mTVm"diG3>хb2z2֡"yCBDmcVr$-oV-n>|kWw]Jj.& Ka\! Ns|*)!rp10h58.epؽG˫_l蔶~ȏk </[>~iqWS¹,JO)|2YS&R.$IyL"HmWY"\)`H-` f.)%/;;hdiIqZƒV \cO-?ICnXɯ4ŷ(1^H^)vW|W~m8;gTG^^oX^]| @jH)ۅ#BS6z} ¼~gYul&2C y8YޖŶ~b| 9UBYjCgמGdyk3|Ew˕(۪|яliɋ7?=ϛ8]Y_Qs%m ^5Gt/9V2lY/;rmAlm{[mü>A^|.9h>_ NE}hm ?|[knU\g䣦-T~u݅v]ϖf ht膪\ EmӃzoOݙonL:ڑl:}|U~36%_rŸqr|kW\]]:o>U}ڥ˶m KtZX |j HG(Ts<j[Gw,8Ѿ/O6/֚:KB5xa1l9a^Ŏ#:c T(Ɨˍt޵k!k;,3Y@ J[h",!"s!`PyӴ/ILZHҐ* )8$ H' 6 ׸̜KcCͿI0h<_lLJ|?tVb☝'t}Mvŧ%>$ hHU%V\s2gOmU;P%@;g$oS[ld6RED[7ne~g8ƲS;XbQڜW0cf?~Bؓ aO3žf)=J'UO(sGPtpW\WZ<\U+gpnJ16(z8,]sOǓd$WȬp$M:Vx,pGL̦CMyՁM#+IoVl``P`Zf#LFHU5dsCjCj%!\u Ws$Z%Wl:r[R2X `JM w(pUE7tǚ{+];!+o8" \UkipU]•ͣ?~]&~6n;{Osus_BviIt֍H|3#a>o//i O;5]b_e}3=/=)'<Ҕ AH v-m+j!ɤ"9 5nN};UO$I- FԙdCG|{:O:I!%K$CI3f$}x7A"oBCap .BAJ0[׏8*Ұ*vEyO$M2pBk1 τ&% A !2ߊ/z~mg|WfMA1J%DN,䜝B"a0w #jF٘'yҁ(~CJD BĴF"mS  HtfUEb Gch %1UDضY?P|qeYd _QH>$P/DX9cIa |qtՉj vqCnڣ!:yȞ!EXAE>Cnڞ~`'jB:1Z2 +>\/ffͻ󙾘& 3|Gd"ߡXfaT(^) g^EL+'NĄ>gg%rXV*#Npip! *.{RF۶_iKiJVg[Ki-Q:öig V@Dj&p::UȟO){!f $m@k tP)#@mfi駺1-ͳ](ݴ=Sl\7Z;鱧Qy{XRT8y=P2$J..kN1P kgmMI <T$CO6ZeRqA&-IF̜؎4f/6.cE"Ueޯ/v(76ˆ?;~=;_q̐)Pf1[R /`"瘞@ zν9u nJ ص6f?Nϻ c,3翮?xaG;H lw*n`l٨v',x*~ꁳY}<¯ε F}0_AIcl\H8 1Hi0*_˜XKdc9+"yLi mcQp~)JCRsv*>NÆ~yru~.>R˖Qrg_ z(u}W,B\Q))B,tL2S$/Pb(Rc*/՗$m R&Ў"HˎBfNGEc92ںTȞ`n[`CuLL."A9.\2Dw D9E(C3syHaH xoW|<١S=g뱃x457[{/xί7\u0*Q\v!{|."`%1kF'6xtvyF=#5)%)\Iub/JvDΆq1ZCCz(]҅e dOx#^h{M,|J&A!s! H)Y#JAi%r)YK3TbVQ}}[i4`?&^KmH'&%&KnR:I]Gy'7(Ng"bwd*W.q2]ϵߓbοKM92k9׷^z#`Tx3&Q5Y\:ܮn*| joqCǔrzU궥UvxFiwp:P;G1}q9bJ}#JH- {7Yw\|ZI%T) HӦ 㒆nAۼ'o&cHy!y<^ 0-$X8OSՖԥ Rd?20ɝ{GPWy>+߮;Z HJ?)ڝ6n[ c <<`nu+b+I6n" qxWXޢwf9O[L'^ѷG5!g4s}vMKvulcu=X.ŎY;?- !QK;7wlsלMWj~۹u{EjSV^zIXߌ)o{Dz2,iB/ێh{;T6H}/9Kӎj]E|n_4noJ. MR[M@2s?W0gs{)<rQsŸU֏|1cӖY?v])Efx@ $8,uR ѣGԤZϣjP]z*/.zwRIO)+).\Yѻ\=$D[qAVUJ$9ƐƄ$D{]`u0h$<8ǒQh= %/1A4C$t2 eKF+!am6 m/ m=t/ wh:|ZjZz39Tm47Ӟ.zCNjj|hz?y]ϱ$+Iww^Oh6NͮaC +Gzǘ򩊎Q;%NSeDkR:w+ui1]Jp+))kyix8փz@[wȹ J%C2*`wEd3 y_&ҜTL6J82Iuy(Y[T7$e>2NgUuhfGo-y\mսs;%+yqj1fL)`P!0BTN)y`u4 %K9YO(ƨ=B4̎Шךge1Ŭ1V,(΁wR2DBx,0+&F,D!rcJΕN49eN+gڛ8tn"|f'n"QY3KjG)ݏ7;[,iFΌDӐi%r% 4-d1Aiǃ܃^:%RjGTrL TXyt|$4X>Z9}wYM2!Ȩ) c祰xNP3ScA&ȑm뷠z;qp:bo{ft=cNq;xn:V}E`OTJXժ B=mqDHw7-=׋zZ/8tx3Kifu[ՍV{x?qa}(ZL⭃xqR*&oD\1@"A|'zhۍYݶݦwXttGMdIוu+j4z2ko]654}v"wú Gtz?ЉՕfYt,Dzs5L[O P|em 2cNn}G/_{R)k%fWUKLX[]#ڏ6Qrko'34Ma FEy[=2oSotdo^t?1xB|;g7|g˧"*BINVg!4-,#9MTQ RZT*i`RrE"u1iDpsy]j^m̬Pdt9QS^qc@+ f.s=$NeD! ,5%1FFFHbi Q %j,ϤxǶL{eKd c__Lۻ_ /iy]]ТS.*׋bV T:?OpF?+SŔಮZ^_}4}}(8Ji*J:n4| n()Gl=-һI;,trC  SXOi`]Jmw?#%')ʺ}*zh(٤!|%JZjNaο&dC/ ikFPzyf3Y]u7 Oc5z?#uշQ `gp=]iNuؽSroLցP!ۉ;E[ u`buwɑM/Oȅn[{|XѼ'xnv޻}t,_ya>?`ΓA>,Ƭ7?sf9]޻_]zhqwKCU:qU^km3zڞJ ӆᜠnku'o-ܯ_O%G9(+LiA72n@e" xSf:>zg״<If&i9d \(T%WL>Ƴ6욀.I$ި7fAXzZ栬#l+Tف-=HS0]2R bpǦ<}HB]J+A9D ZEtR)D2 %tI9',ld5>}Ȟ4AJə!2!G+P(Ƥ}y\GzVgzʑ_e96*Cogzn:䖔[߷xtXlS0X<*~U,wz9֊zRK  sLkTP)jxT#^roPwJhƻxEVg.~5{z^G~tk(bvCB" տ/1YIqn(&8;y7s?r2Dŕ -?x ̾L)ם9w<;[[v#82O?Q$~;JȦninm -okFlmfSp7),4TWZ;G_V/zv|818z[[lwAnkuXiӗ4+AґҰ$dʣ?`l>,jvMۚ^=Qs[Z}aX/p \1;er$xXW18PIAX>߮J2/ ބl@vd;>7^2Xp` mtV6/GBei"f72 L?A*˘UR.$u3l>cV&# 8 ‘*Np"sڠsdda3J#d!L[rJCO7mOʞ,?\ En586{wge^$ZӤ(&pn?nC'y ]N&\NrUv>y4F4p"4¸%1K!0GꡬG 9-<@k:Uf$7`'̲ʣx%|:K㞐hEϜ WM/amSǓO-ہoK.-1MK nk#[B{p}6DH2qżHجs#z/ih,jTa `%Xwyzy~zVt! #6z֑>'ut>KYmTINV-dTTi&u,XURY:=۷iI];Y3zgϮ4śn^ }Ї$aݺӖþ[o|rw%2}\aVS<5l2Jxx_[HJ۽ջsB#a`. )|@u2y{Qه 9D]R5) |Z S"<c"ΙK ZhȱEDſߞ}M0ϋ'lf]~,k 7tgr 3-x}wsLKo5?vϴm7 Ns*\Z}J `\} Kz7 f4 z3ݴnOjzMHr}~5aLHosTM.6B\~Ö7?̇E^37~yE7.Nmc׿2 LG؞a}^yYmfk}VnCfY2m2{s~M'Mil$ۨ%!Y- 'PFFn-E#hܢUo^#hME#hܢ[4rFn-E#hܢ[4rFn-E#h{:ܢ[4rv~/ Fa#h%*E#hܢ[4r*q[4d-E#hܢ[4U i#'EXܢ[4rFn-ŷ}VL<|-ݵjn&_&oAEsE#hܢ[4rFn=SzHi?!Ψ$.5p*CMaMs\tF۔S Vݽr6VcUGTjժ{Dt`qN91Z"$RdtfLy!W'OXcڠ0ώ-;cg]|"6*A̽_?tXCTM@QfL1 Q[@2V "9T`3wnpd,0u)KH]W\^~_q{@%&0Zl2p,y\)3T6j_Co$\\.Up[`JM0O(-E?0IJJ,'F0NG;vhBX3ȱn,'7Gt xfK!dB +;`GkJZgrX":(5"O!*WxW&$1K] -KNVW"V܎ =znJqDZ_9m go6|@MgÚ2Ryzʢx*pgQ&QI4C.(p<žc2_=S?wLNg2!ZD_>\&SJV1&k; jUZa>j~RU>zXcd(I-Ô+Y )0n{MH(|mH2K^DSf;pZ-}7K~,M闣rN=S77x>)oqɴ^ekI&}i"]\Tl%7?`PU|{.IK VW(JGmR`$L 4Cmget1{JBeLL0Rl.E3: ]&#U2V햱Vf I}ma0e4[-\m -닋}eC//nh44/b3BJgr(v &!\CDk `)q{*`HQIDI&jpc(-bn?%jOvq-bդvQTfwF#oe4XKY9Cj@2Υe%#ϡ4xx4{?X=9=2~o?5HkُO0ճ'F0+K?G|>WHzN׻:ui WioI(5yQWSnNF4?;.d$Q"y"x!4I&VBs yŞy1^-Lq:Oܝ;W7Ʊҏoq7,W>s~xJtD֧/o}6ԏG0#R KÓ"i0|;i+ CΙH J2gA%ϥVXrL:F1 1NE6j BrP ;5 S CBm֮j٣,=" @ْGW2gEt&pAw~L=ݲ2n0{'HS)("Ad䜌ˡNudB ; O3b3=[MOpbv7(q e`m6p q]L6oowڋՄJf-pG"0d6GE˦&2B ,tk )NU{{XACDܯwn׾-y|0~j#bؽWӎ=,$Y̔wxnGw-7+5ijgLGJ=2|t[]nxoV]/H<7ѓ߳13.F\Qv>u6昧[ۆ֭O RFh٠fa> o|3j4uaV/ݞj{3aN&5>хcҜjivnqsQ[Ap3UcTFM":gsIW:5G= Ä@c_A-9:#YnWx]y)s2p%F4M@h탳jrM*ŔrX:az}TKYgqj2ՓjʅU/8=zKQiICV3eE#bد>XgVu&>oVڰXk+MW7wo~'$܊>x5=`oQ"鏏ѧv8( ~F #dyndUW8˄wRaߜ׉o~hfA0\E[qB|Bc㻥$PUۭu^\?~4͛UZ-t<9NP矊Rq3yEEB}=mҖֺbcq7Yahā߽#y}7(Ԟ&R kX) >~Cc+o*һUGMh/V AiN=Z(V9+1&'Gb=P& tz|퍥ܳ%L :gH]:c5.\f+wǓVXi$O]WA cr*'K䉅eW rm=[^7AxBPc}ۗc뱎ij{StͰA~c񭊎5Q;R B]1ف + .~єP8bJ%=O#BY =d,JhVP'Vd T'm C6i^ 4r S@ t*l^V/dNW S4wFNZKP .d ϵ8&ܜ`Z?U&i-u]roF1':%W9 _otg]ײ_`g-h͕0+*W*_ZCP}[Ю>^H:zW`G; v#sYM(p&r`4J:KMBBQ, bV$DqŜiQSEd%)ĨunXЀZ1KsDF:\]?KP}ҹf͐P*9g$k-^>Emiiq7>FS@bR"B5&T0cfx^^' ^:+FUIQ#e. X8s"Q(]P`HRgi/ JTr h.,*i"$HA[+#I R3PF6q5 KMB]$/0V> J#H:YO9$l Ldm2, Pe/(O 1ʜ/3t B*!($J9iɹ(^V&%MJwDzDcqMk H䅴I[24(í0=k{ё:FJTB)b Q} HM̀_f@yNy_̀':̀6rHc8895| Jj%`δ䦃#ϑ\@l" @FmiK&nIN@:0 <* ^4*uzp(qo 5@5\"Ӊ"T`u:h!bHg<@D?ŇXLxDI p .*\.qb&xO&%K@,OA,Ỷ1<,DA$TI0K@.Ah,5ADbKT˪d+ t 'ۂ6Kwb[ U`{&ÜV.ݩabLP@s% {(ךY CCak햡K?+͸pt8/~q)d2هvnТS9rQօU pZg"HU1Θ Hq(g;ӯsR1ȳKU)̖oK|D2Úys|ʺKN J\AE%f 4$:6ޔVlS[S wnS\2Uk{NjշNj_3ri$eD$45g*9 NrrZR[=o'HoYҒ}>垿jf᠞!nh:;Ay&Qi%>(;`9Amyi!PNcˠunPw}:Jm=YNhV=7"]D i#ՍW@VaY JWW1Drr7i3v^z[/yGPSS5]jy*p+Al[G|@Wںn&`_+Je5o҃x ,!k՚?k$)kl+/e=n5n{}gm/wѱ|嵑a: Ȑ ]>\_tkپ=K\,10xǶuqk¼[+ͪ>mn,meic ꯇKK鯗]y -}ﯗd ;mnDe@t8PFJ ɟG+缾ݳ#f֜$|]vZߚUA{}ʠS2[2^Mao'+a;nAeYQ@@Frl'lK q!5kWC?ϼ5My0Qn}8o٫A޴\s9 Sf0j$7.(]cpOQ;S؂6 3!9q\[ŸvָXm6j!G];oRCBNFG6cق1K(Ê_o|z+X0jA1܁(w|]Zq[u]d4"Jv>0`)O9 J- 4"簞[vGtaV3D)S!˃:_|ėE bȋuǖal G\g\Z7tGV1- Y}/+*AՔrAi^RTt6blȢMX m z2(ʡ7ܕAydS>.H10'؁qoktuf0WKiRTL`6(: MiNsVF$Λ۞2!]:nv|7]?;2c)(x"(E"a@Y`Ȣ] 4HAjTq>:LAZ <{g#~V0wTL?KQ `*u &p`Ej>dT {=p) r6l#A"\=+!!訽Ŋ2Y7+qG:` fYR`(ِ/uu `[UW ^r[gͺT"f_|#GAcI5NH!1T(`5B{SDA e͜`3DYk| $n)&}EՊX{ԙcw=nv;vJb6(!8Y|\1fPTԃF!NH2]}~q۝> [󃳣.@;x2Z$nm߹qe݃ꐅ yd6Q HjmlAhqUVacJF9JP0\qhh}6`Mg]o.Z3\\_-zLpf`H]nn8c`3=+[{ 08 `Z} \;Z,fw:5WZsLڌQg5r@n1fr~1g%з5H#t5npJH/{)u-QUFeEt:%\mEBv0X $SSAv =>2zwuz\ ;fc_X!Ŀn?q xMRP'wCȟu7uA^jH c,1@Z*(.FR<z됪 lU0?<6NmT11mA rc[ C5kCVVmd5j>t&d2Վ *W`Z@׶q~5801^_bk΁ >b1=ڡN|HDV!iVkMyQ6.rċQ+Bj q#0[J$=k9o\jR- Z-`XKEt,*f-IgJ%T$b~Be V35w OW,,$; 6[x#_^r}}{?1>MQ>u1 aGw.Ξѵ}XJMD8%S+2'~Vڛ͇?<4ʬpup^E B6-G ,F ^ N@G 4N@%(D $J QH@%(D $J QH@%(D $J QH@%(D $J QH@%(D $J QH@%(D $JUu\r@|T-F @@)JT!-J QH@%(D $J QH@%(D $J QH@%(D $J QH@%(D $J QH@%(D $J QH@%*P$%SX+|%xJ ^@ǨJZQH@%(D $J QH@%(D $J QH@%(D $J QH@%(D $J QH@%(D $J QH@%(W Ŵ %%f)J F'@W (D $J QH@%(D $J QH@%(D $J QH@%(D $J QH@%(D $J QH@%(D $J QH@ǣzsk|ro5^mﻛ7|]{o7}'D ViZR #6,G#\`^IpKd--pR+u*,INWN}|gWlή祫PtXt@WQSyC]]1`CW 7:eBWGHWƦLvla3r7<;W;Ke^l (w9<9ۯ8yy}ՆѼl>&'mOT:g )w7iӂh^M3\KKiF̡4$4}4+vJ/'[ ]1ZePBWGHWD> b pZ̚رt(:Bb9Z]q1tpbP%I1xt fI p+R hvNWc⢸ ` +R~&O+FUD `Kv1tpb 2tO#fJ-i#+S 2ڏ0tbQ?]_zgJ> Os5y{t`t@WIS jItŀ_ ]1@K+FPFMBWGHWIVi7?>'3*LĽAVv?[n/.W7V~rZu{/Ԗ+8~|u՞@֧$cb _C]07חl3im+SthTri<}ʍŲ{c"xu}s{Qoζ7S?O70'f 4ݾ|q9r_Τ=a_]߾ĭM-#^Mew7|-Vmכ_U/Nn~x?mΦc;|W4t-ejS ɮ'nt$3uv~9ڒOXsLX,1`knt~/S4&*Mŕ SȔf9 qy5,PLRnc~yj“(<SV_Wwtz?Wld˧6]6g߬q=+R a٘]$*q'cN)!шTio4cmZ=;]ЂӋYa>.eAt "@锓#\Q!-pZ]nR~)th tutE_]7ˡ+kRzst(:Br;$`CW Wۥ=Q:B1&kDW 8-g3nE]1ZcrwGU01(%/i7W7?v"aě+x/O~/ًv^~V>E/?Q疧n(<\ )],s-3|nQ|Tz,锴}k~~f8lM0 Ŵ7ܝ],IvDI o*QU)+*FnrC}UOVkO_z}PHWW+R90&:IԔ# jz6'w^^p/.HG/0ٴ4偟iΧ?])!g^(`BEfW`\Jur!=XѢ`h8z15oO4Ncvcwyض|s+3ݴ:dPԋrp l{>Obj g>haBRְd]e(j#DUy3Ql7+n+P)rϷwz Cc2QL!bbg2f5XĨdCqD%Q=S`v6CՎ zS2NwDMI6!5v=']0i('m//v ~ ^0Xu4iNJ0MdpZaa 4bv҈=܄^LQd_JmWW3zY'_[%KJftjS ]M~__Їڤ 7kOmzrfSzMIXXi|6n^)'R`*MaO۔u5_eC$ * ӀYCU!L":ݑ(B*ДO9!x`4ʼn '<[tÎQrh6J ',YO lA4d)2 K焙!s6Z+U\Rh"҃1Wx2(1zrYJf.s!e:3Mh`Gy$qFф87F?~G"A{~<%SO6=~AgI(HLOדy`Zm_w۾9/ ~ Tޡ@cҊX(:FX#VY4)К`B(1J: tV%w1a׽Td.!\*08%f2=91sTⷻfQPX[Mco4mÇ~ox. }?k#?|~'y"`2s-5Oќ$P L2,e:Ǹ(F,%I c$ "qphs&HohH0a lu0r4Y⩔%OJI90:һ-ɤ09(@lzad,LGPݭ' ItoJwjqݪ}v#߸Rb"Zj.`:TσY_CIȃj4<j}FizFYIb$y$,w ĆKnʭ&`V<4f߿S[5҉ )kIRXrvp,8i5(irҼl$iooRDZ\.q2]ϵ{ύsuW.~=Tra}5[~'9 TOOu|yᏟқ" _ 5[.azzSp3[zzUmEvxzyvtp:POGS?cY*("U 4dWC#ʦ*”լ^'|慮š)yחO&1n}Wwi4i^kUFS4J֫,0ɝGGPW|Zo׌a#v%)=6nNc <=`zpQ}lGb63&r@.m;-QՇ-gCЂMpuj>t],&t"n?D-k}wj6v,WVmg=N}h3jNS޶bu4q{݁m)qO;+s9Ҏv)Ӯ9P *欴)D$R\z%%M]P8aEޜ{$xȑ7=m?ag̘C3 ,5*ᝃC Χ kb^/F@oAUmn$mp+>zQ*G?>:8B'=Or_p@V=`VHa9l܂VXaTaŠ 88ШԲC3h\8 zNٓq%|#kzyRcR%'L)d k0Nӗ87vyH,w-T{*>Su;ޯC Zw>0 h}?/Wgi[-f=kb-"zl hC6/?_Ң/6M^9}i(tbu$. J &q|_IypIK&߫ikҋF2$IV20.=kV6qtIBʪg$14F=mӇCs3Ȼaa:HvoB^cZkw^W7~8&ί;Tdl?d~f}{E&U(o4dJχ%sw˗.$g\-n.&-\Ë1JՆTYceOw+u)\J+U.#*kyn=pޭǻx։sA2(2I,L6E01Y$:]Yp't9R ',4:M23Ͷ&Ru8;F\tٔ@:#$𧦙lpNinP.uO[T7$e>3N;gUukfgo-˸νslejy2se"\MKB<#J+$ѩ>'lؼ%ougd߬ۧq^cґI-ϥJˁtͳ1ŬH) Źg!~&!<J+ '!qsl gdN-Ċ& [t\̦p7̴"ܝO:ܠE)QEjNS%~D9QDn> sb1:pf%>k>x',I' QҎ/ƾM8/1>%Z)~W=K(P$1`I]]]:a\6 XD  YOq0ȚKׂ0'2E44G̴v%e7dń61Ȋ mn6:Cwy5m05 s+'nLg;W"c%pJ^kfE4 h0D.2Tőnn#1CgxV\w g5/nuY,/G.o[eXёbа&@ZrT5'ವDB"b Θ Pa0@t8JCkwݳ҃k EB!{|rD3oAiå&*Nа iEϖ]*9jMq\yK< \ib~ݼfs%gz}{5ߺx˖nuބag_iTCF!f0mK6iPmfBݑ8~$>ޛ,q5Bh/AY[PTZ G@dĵIw(M K$Y9 I)4 -Cɹ9*S\Xզ9ͩ#"nwl-f0އ^{JB<3(7kEAQ,:p%$\JS\jMBLə,F fU5T$52J1hښ:2W{S ^r*ybΪ9;qXY}OZ'v}}:EJ.+ZrKfL8+cCY"rx*e, ToipC.q ;enp-p48`jXJ2f*JQxke"$ D6 CsY&Ҵ͑89R?%-׸|QD\Bbt2xB G%IȜcFDFrv=$.g TJAirV1)XHڀF8ddNGE^ N;Aܗ:W?n  ^KL!PD."V3a9!E쇺 ɇ#m+۹Hu>9bsJFD $'ZQ Q8R4Vq5%FtyQMJX όDV-AOt]kKrsm68ϳ6yupe}>?;KΎqmcGG M .w gϣqۏ3 8q*ʩD %WFG wݹxS̥6nR@6F ^S%qiqhx.>(/)Gƿ$rUQ FbRSr'?ӓA;S*LǞ 1GCe::+[ jڞ?WIW%|hs藺 v!P)t]>qGLaTѧE}{6Dk}Uu61ٛ^M;t9=ǖGneo2eꂖ(-EͰ2|3m^%#3׽4egxOp ;E}NM9iqFr矽ϯĢ,6!-zoTQrK%:r;WgȎAqt݋?_{~{uwO('/oq 0"h! jm~] V߿ M5퍛i4р/.gMvyAI2-Z@)~m7 ܽdՉG>hw20~e~YiÏ^QQ9Qt2T"D7b0#Aufc,?bom'b='#1&0`4#.ɄjH:O.!2Jld Źhu66E^Y;dpO-)‰KA) ֢>I HM"ʠR^qBeOu{ș v%vcOklul~7W4m'⵻V֛xl񽵲UdXڴnlezye(rq}OjqU]g`z+^CQǟ\yu0k<rX 7֙Ƥ ם|<?SoGEg 8,?Ú਱)H9w%X#Q Ɋm V{dS[~~՘jLVcLsZwq߳LmnG147wq.X19B,Jg^ŸNzk l =|aϚ{+T.@s<3PG]ZyQ^1)pJEJ[k4j+BPƹIBl3 L$x$h}bΩC*.9.W Yإ{;^3Jz}}Ȗsr[2߽ gKЇ 7}p;.T_)$”\6DeJĔN8S&&ECOtLd% $J <XC#w&@S Rpc)ffL9 X',ޓn<]lc* |q}W,[aj-÷PoKP/瘕CUv.xzm>@$dPa,eeV&ZڌmF] :k NJ(].F#{IF31.oH]սcA]x۴.}{Ѧ~5 >Z% LKT܌GUߤ!*\%JaV09(xQJZ!1śM\J\e7}}\gXќFc$f0uKHW/o4Mq09RA0LJa O1\s#SM.TF x%ɞsեfkhh[fK*J= k7A\%HDz`6H rs0m/$Xvker"W,/{[q3H\Ur/ ͟w\\e*gXW_&S?jq!7M!#V.GI3oN['AAıDŽ#ϊ~oXp4QۏgDIi]#և fz愣Gb 퍘147/3_ĔV{W`iF\erqJw]\e*iW(1L=WH07*+fUdURV\}JR#q%{#2L싸1d*e ~JQl2blW\7G Z%w]\e*l(3|'dF'+dċ9UUzb{-}9et[Fnx Z/Kp(sG!67ev{; ZՁ⯿Ϥ~x`3p$Ia,u1f"{A2%RH@1ʄUl(ZC`nY:2" ן^j2]vsnX.7oH?;xi^{$Q<E][Ue}P=r! 'Tक3d@AVhHLZը^{yS܈whP$*\ k{~ͨ?E ^^,X),{)@rjɍ}ǯ~C .ZP]f|*}iQ:IP_۳g{qWCzz4Y`8lamz&vtU;qNRyN5,Q*IuScUmp#+-g!H"yp7y P%i>IF풳6'lTyvșVv-%LdG/tDRREUI>崻sYVg.s.{gBuߤ3rݷ=,J$h(GG&( hd$-$ qڕ<텋K\m{H? GBOW_/xLrE{9!: YXmPxesjr7mdFlz}NzDQ|X)6f_Z:b *?y㨚+tx7y:+z应>dK:5o]峃IZ=%mjWՐMF7o۞?ZWԺi]-ZsR|߽MԲՒ[vΗz^kfVOwt!6> ut֭ϗI}i;Y2Pv}B}{zO\ifVşZ H`q\Yfs,RQ:%:ݩ'Ҩ09]$o{G.mVV֧u8>}r'KiH<.yiwfr˼t >x/A9sr V(%c \0+f ])1T6no$o_9=m]7a^ȄƠ}CA`v\[ ,rڃ  aF3鲥pj͌Nde $ ;-Ou~='uF+:}ga<{XHVOT[/tDQZD4Rg )c<)$ZsE 76w/%&@ -ѲuF΁.km"UͰP2 m)ͱLɀVqfۇQiϵ[D4 CZ%^H.-dX0>N$қKsi5qДƧH_q n).8o2 θdhE%ÑtWGQjlQ,9 zcɻĬ&m' !7^Ѻ2q49W_oSS8ZgkF4-o ?|7r6VA0ʚR(Ԋ\wjRf 'NiHU&@lizh^NEœ/>Ì&}MxtrX[iov92L'?]̛CO~׋;N?NmKh pHgmÈa06ɴOe.>V ]ߌ9:wLuTvdۨmsNpR]JFH,/4AQC~!&өlGx&ǃ0:?%v>x\Oi<J~4 B<;COZu54Z8dh嬗 ~Rt9qqcvF2IӨ\էy Ilh˹ͮI'qW_!I!z]4TU4apbėrQ_v|Мd"̦% r+0lݾi-n#7i HB,B* ShA%KP7QFzngJU|Ek pz'rH>G 8GfM6B63m(Zʠ%cq?LNGo: ?;N`C\x͙?0H1';>Xrچ<*^`X5O'uƅOn;cšK}T1%oI,E-re[dozyH=;p<#xug@eh6oW](oONn&Fh2L~﹮dRJĠ&N,8QcNIqdUj@VENI" LN&zRdggmF3`$Ddq HQi@V_VbE-G~Ō~dTF1y@h,2 P Ƽ6xCW E%G ZknH| *ODU}vF+mWB XWI3OWvd- F 6CZl}KZl\[[lGr}Pq!GJgu=[5%]EG $@l`&p`8aLU?ݯ]pu:?{7['>]WDIRhiŒ7LIIwDbF$06G5ibpir@ir;$_,D \J*GqisLD`'daRk92,{~ )[2,Fk2epVy]6*j1;F= `q+n((90˜D5+ V#Y0&|bV/K:DcpR`u%I>gV{ [eRɷbr<@HғA J^Eo J̽9w0K6+iU:qZO-?)?جOCsn.LlmZ~QCшH(+o8[BpԈOE#2H3E.E 5Y$1m$II6k"s3>@ E'M%a Sqi3 A֮>Z5b.yҋmc!:qɱzW֋0Q/O Ge)TKZJy%AYV"Jbz!`~jqPisnvupl;e ~w3(;я7E?E?U~~qEW*%w־&LevbIMA.Ά$xTq-eTˎtY35^Y.pLd %3IAЍds Y9uSF6: ,0# X/-,ǔJJxD,@9w .p`xlHoh $F9`ZA)g$"7L٠*a#T3x[. %R(p), iIIw +k#h,hYm4V#1CQR#)1y 2$PTIHjB,ɒty{H Mҁ]-C^rX2uH)DfLܓߘ+yd)"筺UEr2d}Ò1e s S0(sT@b)`^yB2besFKAUcGR(CE~߷=9B[-x3ht~#w b,_2o;C .4FJgH zMeg:^ųWt/nYѡFN右K*ݏܽd?x?jOݏJ/\iCrR ]V9;?[oٳiI|CX?A:P$ҋ>V7ҹ$3>nwó~&5 v5m^uoh4niuﲽEm6ݶ }gsZ;;4/_tv5^C^]FwtOV˫G^00Mz~D] 9d8'28DyXۈTw-Pv-iZ$& 1 SGKҙEztDeF 3rhhBB\fF)ArBht!9:WV` * #BndA{cbcJ9#bPuv9j|a K9iAIY%J #O`뤸fZD 1*< C]c$O\J2Mg,f.5uS@ŸWB+ V[0&#(tvUG7L ~pWc=N.l q#@ɇG\1>B!DzPG(Tq)#p>mO"~4ꪐkcQWD-=tuUbTWOQ]Iݽ=H)~\jYyf7.M{hE*ܾ0BLM23_ftFM;@7O 6gZxȑ_mQ|f3ٝff $05%Ep%[-R˖fX"dXU\б=T;~Ԟ}* /0``tSN(]ido1_Z`Mog5Un=*wI!cSY>l$90NKF;]EPC^#  9ePxipH).•o` zsI/VMݴ0a0fG%H+*Yap#:xQJZN!rq6 ^,R\D OL3Eb"C0k%*VE6A UNBakͬ8SqB$XPz2Lzr-djk!SXZ] zb} m5>qUp*Sih#Cq%2OH\!ןғԲ׮2\4;WhuJJ cW\OE\ej3#4ߣR:$X9qɕ']ej:vqT*q=+HӨG-aI𫜰fBDW܍-˨O_nyVSlc k,WkϾwg|.wW#4FT 2Z o^ I$V2R+ʘI #5dKБ&-*~T je`6G;ۡ]j?Szo;gOdԚg)HۧiХD7B]ݗ 5}`j!rP+2Mrɗ||IMbSv퓫AIG]h"uqf I ShaiT:BvNkt#MXp\; h@37F52<7ch@:ATA҇*&hyh(&7Τ&b% zV3Dmlƀ@80-G2kꟷ|PYI7yi(맮|7bWӕd 7\RTrmˈ2 45 LR$wג,leNwǾ=Mr^lQjAMF;gI]tV跍}L C1T}3MٚR^`8%sYYHR  A3rx4a w9rcfx w~. eѨӆ\Pn lu"a&Iu h.ADwV1H,*j.x뜡2ԂQM6ʱ9k#gKZbSC %',U"< YB8e2 8{tK=dW}YQyMHIr|zS*= w AW0*#!nFhgT^ƮpzP#}3";!t<ee>ۅH^TGmŏL2[wF/fA<52A~R:WP.vRmq(piǓEZ Y]H_'tKR%̨(tdzPr",g٠y}Cbd= URe .[uSFR<)rـzO?pUŮޚkyUhƶ|]oCmh&+[48퍾:˕͸,;e rJ϶&r-i>nul֬o:zq:Ob[| Gov9v1y{}8цF7ɻ~6,ܺ}Aw/on|E oawcކQ o}''_Jf/ɶ^=ggn^btB]ޚ0N;7 2OCv-aO3yЖDUݩm|*]nM 8Uh/@Y[PPZ G@d5c/\ٟ8K:&2) ;!\yH<9"'^%u j1`)kLHڅhp`2Qf$< D[W9[: œ=vnD<v9_N]5Cj_NYRoL?]EyEV 4RMZ3Frf,QxfC;* Km`u4[mͤLh)J&7Uּk#g^g}TUjdWc9WSڮh,g,Qڟ4LdrOe;@R4p#0fXaQ7\X\N!uF J=>XD; D)Q Mh0JъФ.\q/ yQK9s2Y7UKϏ%yz:m[eLUӻ8Ir1ա9@ԨI(H,4M5ͽkF&6Q"hku9>+wz+…sa=76p*)cP IHZĜS?SR@ͷ"(% }L7"z|&ZoTdS_| m.Fbzy>T>yTLCP knnSQ7ṇr|R^|a}O}^{пZx ɱy :奊`)-gXx7ʮ>F;AZR0IC21)#knuŀƨ}|Z8;9\7hBۘ0d0i3{_IHϾWD`YgW8 g((|o=5MމjpDd˛\x´ B'q1KVKRs A*iTF!AH1ѢV72B)jcAwxpV[8QȄQx9i ]9[ sizD\lX̔t ϙ1ɫK_Ś$hsӳ"A˲ m{μmRm*re,%G 1:ro8Lhr &D]&!c51"qT-ތPոwcg,o3]*%x4^%)XHڀF+J$(;@|t=v? {{qXn٬q/Sx<@kEӖ !!$4oY?jm6F$˥S"6B$j-$9Zg 80@bys捴 41,5˻zmPZxfq۷=b{C¬9myK W5xwN ˒M&Nc^l8 ! ; %qkqfGx_FcgkZh*yDg#1 q59+[ǟyO>ʏoC -3NB|\7ߦRz^`ZVoaMRІ x|U4ۭٵ.MT>f|9b\/^v..'s.8GZ~qu~7nAZP=]Wu*Yd,2{ܴ!|@Sd~u6ѣ>;םDV^gjxީ;Nezܱ@[XV?h3Oz~ШdMyIcqwȎBqt??pޟS~O8sD% joAGt-u5lӵs ߢ_rEj;~KcGn8cgso׃Sx䋖S-NVV}m~=IZ~DIQ%w C)B4/ Fs9R/ P/F.#1rX8y`sy6c2 F3 % YN iq&jXVŹhlv6D^QoGp 'Gmvg06T'.4Xbtb&" 7(FS{ =͝N-w:8ɾ23K~.>h131b7w :P AMX҃̑|PsBq!++"@ʙ:罊Em 'AQqʚD6fE!ipTbv5I}enPcOxEwTB6B,Qn<虴#I$'J'JɥT@)32S)tJR)б=<sK$"/:28Zl$Rp]H4B*ryAƐ{Q@ˌ4ɣQ8SȄ>=@R> ^.eVr]aL!ڈû;qQa8/qX@'.Q‚GŊiQ"p%'}T, O+_I|#WPDd̰ [IA[T*"TzȺFX })tLq*v9sEU5! =lT:UI}[\;,_ ̇۝쇙 d#y%Yb[c;,+tIlfWwW*֋V4>yE:RLAu6ة}ǀ!%ƢK={h;)H*ؘHB"jNVJAPhH& Ay>KH6FKR)2%E/gdVF "Vj˳9C Mrx}n8OHbLcT(>*]ഏ#L*<&2 >AU {șQrInw$X~x PGfPK6 :eN]T&<IrD h%J0#gpFil 5^lhdN W\3౱5#@9^܂=ʧKR" !(%K2m]**0`k4ZB標Gh)_ oz&u50.8$G" FqRA-Zvh'Ijz-AT}Hbm^SCHq)E*E =DB0*&cPx+I n",5aH~"YGaWemK$?%*0m0C]vCf,r·I?@<BI_`!a%a id=G;cϯ"8H鐢t.8A/uFS[m| %+z QCO cɾ`t&TOs i(iye%$9hBQQw D[;XŶq cUT=S$)&(A*"uÕf<߿#e!:qɱzQ6֋0Q/gE bմ*j$-jdEȫ]ߵnԋOӊ;L{v]zGrkA;p#>C)hճx_#m?(jxFKAh?!Gz_κ__R˷-Z.Nyt6*Uo3F(>[=G7O=u3PqϠ}L T$zeVE9IQJY:6፴3"FM Imt4mH%#&c:R*]T.. &QIS@zfa!B(7vr4#E Bϫ~QRu#q6;zqsN_bqwȺC|nOZK45$)ɬHV+XF8YI4Bt÷Y¥L`"!{5ΧB"v?*h(%es RŃO֢,(.eTNZ7hF c_A'?;bkJ/5t*){njv~f6|j+ҮTU(Q@A'yB, U58N*;!XTix՛4g<_)%D*}IĜMb/JH.Pr^`p=YwG3k/^Sei &iHBU@^I3J}3X~ކ;dP8T0QE-i|@X#lZ(*A]m^}j\(],hETUK)}qqBoe!eؤjWe%K?*{I FҚSwk͛0]Og|O/Dj~'z?=ՓY~Wg)/NgY5Ӫgd4_51N^9^ sLYZ^s_p3 DS-ӗOPY/>uc_.@M϶5>zBW4ɿUP]0+./LM9"l5=;|bJ.;ԎS?O*V 8fu LW~?U!Zߕ@9YjҪWs}L| @4/'RN]I?:M%F5cF9d?v̪ ^z6)0uP r}'8:rGrJ/w^N/Ngi{tJs~Cz#8,x:? _,rzV׭}ѿi'ozl-_}rFWLW3L<{|6 əK:@r]%\&sbHPs>0]5;vܷoPכpf:(桛W&R|5%W >cD=fn-R9ѿ&&-&!\9C7/lE+5 44{͝+;xx0]2x.y1n^Ddsە#.?|+ U|L_K#-MXwG<Ͷs53iw_VM.f3yCf@6f=tFoo>1Im֎0{?/v=I fƺz/lIՐ? W{.ul.OhiԹVD~cD']9tWճۃͺO3.RɊjpeTgѦsKYmV|ұuhA7x]\ cjU.`0HvPJz(|XFDUT[fqlqG;mqu\O⎡TJ|OH{)ͳ;j7Ő͉1% R7{mff f@:rNdB62P+r/3OO$a+ATs\ 1&O7DIB_? *`i]9zyHmWηPi~~>X*^j}. ,~a/#xv@̈Wslwga;|(EݛN@Vik׷Щ}ANJF 4@VY= duт„"X*A6ErgsmqA d$'tXY)*k jqq/n/V_tdԯ3Zf.Bf@ )S|"tm=fN6 FL7\Z%QŲvݠ! PHSMT-Wo3ٮhMsNVmq[udU`jbJBRPN-ק]vp5cp캣hdxwۧq\XPQT`pDJJ \Ifi< K쯻h̛5B`bFNJБt'/س J:medJ)YF` .KC=b>0wZVkz.]):g-^dwjZxYlmKђf=>%ذh-{=KԂ{XSzQ+Lz2lzV7RlbOWgel*ԅmK2xvxm27d1۴}]k6hwu:X>|:b|f4Y`]P볃4޳MK~z>XοK*ܬ`zq>@k|Bv#j6}\ V _>\c7`lcT)0ɛk4/S]V[J7Ól܇`17=J{T4 Kp1mˊIih/?M%=gge陿LO}Dݫ_nٌ^ִtzB QB̦y!NҼZ7/Ry9Em\NBt6":\Z ]t:A\eyGW$W\pEj:Hequ:RJ׃|+ 2\TZq?gH s"NWqrPѐ HWֲT qee+K+R J{\ ΊW$Bg+R+MqE*np%Y3cW~aJu\\S]Si:]=v-zXY!2 <\\pEj:H%W+aL  ަ9Ł3n,VMz1 e~t'XI ՚Nhi}]GdO 3sU<+s8o&o.Sx[x+A@b1W{ A-W%zNJ\2c=G 5O󷣫+k׺f[8lZDӳeQJ- ZY8oc/K^)c61sqZ'a}XJaU PPx7{)h>v;+7QpW$ؙlpr ":A\)Ȍp>P`"z\ 47 U >qE*qu2I3 )U<\Z]Ǖ:A\Y!% W(d>+W:?&~) W$8P;q%24 4&LZyUJ{L pHel:V=*7TBSĕW$]\'uWIJU: \g+O̎*:Hqu2K CF} pUwWRW'+3.#\`.ArU6Bw޻"]Qp唕" v"\\`.\ZTw+uϢWLs县W+{Yvrlֈ#N%tl4j+jע W(؁W$W\pEj:Hs=NW@5ownܐs36gN@m%cdh'y)6jk/v*NJqksg cd+E."VuW2q%@Ix C.9aOd46#L`l6FN\0-kiTA>ALid6BJe+Tkj\+k{z\)Λ vW(XI H:u\J{q9h3 6, H-Tjqewm[@rm6uWkz\Y:(ߵ+AjupJz\ PLphV Plpj5t~4z\}72,zs΄~|\ıJIv|12t:V_E _LcmWL_R6-_|ѿSvIm'oPK$DN y=a5O:dEr_fn_fARvWorsye-053us͓lTL+}YSRAWA8xˤ $g&J0e|yÊy'8Tv<1=K E@).T>HSdRCp]A͊ 笵ei?aPm1Y=2Xo:[p:4_5DLk7ooL9鯋(]mOM{G^9[̈́s̫EK?~)ߗrWgs4<IQ~vy1OxQ' ɇ.x:,*5٦ RA|L 8cr>r:}_BA]t7"07+=e">r#0 Cg@^Bz9=h17>Fէfxo86 .=Cwxl}J{l1uJ'A{|B/ 3]]i{{C6'ܸFnV0հ^7;n9:S̠&ųwW' F9T!5vǢw\`QCۄH_zR< nߍ\ƍf/äY #jqwunP~\sc-j:ےX_-$ljờmҭ{ƧXq{N!-2|q[}9M>4oʟulݺwWObՋE] ȃ3P\Ma;i(a<Erk\Rպˏ{1ilw+o0.c]N汴>mKδze( +xbLL 2-Bw)gwDܼ/WZWa=5JU9 }U˫nͮq D/:7>7^4^(]#Ɗr1xܻn>XD7c*u>WOp|(7|9^ V(e}€ U2&_h+.DV&n#Ӂ__1:y?u"kHrSWt64j}G}1}r0;ΖOL@tBk+'JtʗU"vF Ž&Iq z-Kg3ӡ0Gc6Y(X04S{[LIgş4Sh0>PeLSf,o%fEhJ#KU(Io[2]?~AС_p_-3Mm=ڕ&*Qy*h2o2S\*UYuqLA*+1ܧą`QYZRJtLB9xUt,L=Cvcwk4~d lR24SoKc}ٟ#'OԩO\l)[OlY|1i4sB1g ^YE\L2I3^)#SdNzg^)Ɨ7la|=_4akw14F*@07z~d2rvsZi34/{i#cEɬMr:OI؇=Loǿ~ҨvCdeccǿ})o^p{_|fX$,a]C0O'];tkuMVnӵyիE׼~?Ub;ϧV8xqz8 }}zr2-!m5"Uzq6yRJ\5юWF Beb~#iw9b: %Fm7Gĝl$x\_zH묜t(3!Lf&U( I>qbLBŅm^lXC^jwh {h :1{GdfTVe0 qd# t{:Utiq:}v:uL\Zg`[2'~Kv29=Qpᕋ.Ъy M p(0Y:)]$TOO~8#%a afU&42 sWQދexp;^<Ń\eغH 6XH,ː-cMT(hϙfvygΝZL6LJtNK/uA뛢:8I9brdIH< BiɹX6\d> 92AEW* S&W5x&[YܺuF3]Ns]Lu(n+]x%;-T V%#*>Qjy|e'DfX3!$:MY[$ePѼH(*SJ !Z\b! @yZ3EDTd^UfJՆ},C)mM[ s+lPݜ&'O?&{N=;zv$"shh*gF[Wg ŗx]/!ԣtgP .?F%ߘh/O'۞i==BɇqcI[]E]4ӑHRbsN"U2HB5@_;lpn7 R?TɠA0/8ŤvX-(❒|h` Yd eTdFMDRl󀔄BfHNphu4Zj4)+2P`cM֫Xs`zJVU<^.v HXiצsm21yVV2mJj`މzVm8[eq]'rN2htKəE&E Zx%}&Նs;2*հd +/c cbʵ-N 3__lo_jqj-dΦbȉ.xM jh2v̺Lqd)c Նs;b8nդcWP3'y6 Gs98F$ډ1i=FJ/65Wì!dHM E2511.O-Y@*#YC8L PY&;agF1@ F%x 5*:*%Bt1I/pe-׽W^;2mG Pm)m  + ^xu9!|9+8ڬAr%+ A9ιӖ (56LIÁp,0~wWCt5 tMuv~8Gs~d50T4c'RViҫRn1Nޤo szXv000Ӄ? @Řÿ?M>29۷?.Rz퇟Dߞ2&p<ߧŇ֕qma|H(K04MAЋ b8:|}:+jIv8Tǡ_~?9}KDlH\\[O;=)t{}bBGI<:]i:d=cI^œn!U)f5+86-~ҪhUiqhc#?ƥS`puj=18 kOUDmѭ:\)Zy;ߣjo2{~8o,g㸥xg4aMG,c(3}-?մ5? kv]moIr+?%Ȏer%E6Pf1ۑ}U )J5(X;~j}+&Xo<٤nt &gȮi>BW'wjG~ٿ9x{qoon=|P$/R m+qN~<e]'RڿhI Tea7 x~/>$#{(5 te4 C·͝X/}g^2xny演I:݋NKۈ/*2OۄK`DՌ~~~{?K#G [}n9vS? SOr 49'&7tW7|W o읛7yӍt^OLRlqb`V{eg&m'}uwl[Ր?{>plx0Yq+T"D[e:Fvar `EZ!.`w% {>pAq+>i׋4!߀ܿRX[@$(zE`CKDc cYt.xA鍢*٫:iƮܞzJýWi=+]HW(mwp'Nd?|owuh`^ԡk>j"rZAz0 , H01:"rF$z _|F[ HeaHfΩ,ihɪ2vb&d+bVsmTcx_EsA?7z=K>L Zo^})%^[n?I͠r<]C*q޴}ϛR|5y?_)?|z>~ O5TjGgߟ~O<.PEK)kCoL,KU]9K<[d]|׃8~ЦuSf2J߭оm8ܶs?w&n &ﵺv7z7{}t&l[|`P.jv6;~aҝ]ڰIa7w /ON'WЮlÕ߽KoTntNNwx4xVl&B #iIVSڤRM n҄ؤ;2 y5vLJjt5t^YEB1(5@ɔR)1-@AèfriQ8dl'KàbQRl%b# hm>~kJbbxI5AIj,opR*kw5A5 &gʏF|* EHyT֋DN[Mb8}*nk3-R,M<|7O\~<_KX@CM@`ӈ(] 4Pw M{2)m[(o^Q& &XדU¼L$n^#سz16 jO^5PQ a},gQVnbݟh/֖*)$WKER6(X}VxჩYBJ.1n@˙rT_`%4"uɸ=+J*P%;k4rF( %%gD) ae &%lsTICǖAz,nM_4LZ\XZ_Lyөium^h> ՙ<ͩ +xܩO.>̘MBRA3K&D0֠)۪^z*b8h/ m<:%CɍUQ0O SEŒdFQǺŞ2lkV} t٥9ZT[;λLcwGV-$-Ș.h~:#J d|~[l9=iZqOvW>{ v{p;ؑ 54<'yUͮu{9Ub6zsq.|l[pCvU[xic KFemNVwsjn Ro2o,~T٢VmtG]Iƹ]c\~~zi=o󬐷 QO7tmtwyOWqe{/Ys1EmʇLl[n|'LFh|w7կCuîa~󻪮>|e*;|sZ]yu&~ **٦dbQڅyk&WU'Yme{*,vY{P* BJDWb.Z(8eȍ*ZG-6EPii HU#G#g۾@o_I C돷mvֺmgqf:^ DO 0ORr,:FDFdMS[)h!DIi$mFH@YiM$O%Q XH"z,r6KZq/[wV\m1lW\/p ( ^M+x-!gLnaJo "W*FuUɵ&BRaUA]} 53W*ZQ uR u\z-Ax1Ճ½zҿzFzK]+PW[/,jme| hʥMsQ&&aN-P|!=6Vh\9qÔӛ"<4y ՆYRlXMHG -D]aoȊ["=8z+^{gE<_*I7 }bV\Ժ Xn-hX% V$7T?1we@qX!5?u'=V|;|Dʻ;D{#j$/>Rz@l@mGBH"&_6SN˿#I~a:,cG?g,-de!B Pg_ٝHF伲:`Q J!ʕ>2P B$xoFG!$5('ʼnd)wv&T/th7!.EVz44U-(*XH$0D!NѶ`߇Y;t&~ѤUjn7FYPwх>Hh>Df^Z@uP< *}t̴*M!RrZ BK(Id'_tyQ{YHӴ£j4ZIP*Dw[6R饥GoY^QE=or@f-ym:hIYd`-6v+@yS>vy :LZ1A[.!lonihVTGL̴P٭HEŨն[ۤYkT}jQ)o=4z; ƤE_ 2q89̈}bpPaRD,6vP.oQnDВCqnQD{=DKml  J`١ 5@R쐥-HOOdPJwVKl+dOk ="($'M.1ҠN. )E =]шw&]t{`AA(Ջ6Be/nvjzr\lP1 4( =lV~nmPbQ7˚8-G>ܜ(1ͦjxu=EOϜıc!pU(䖪@ 3UP_HmKT@V tN + N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@:^'u@֡$ U`&K; b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N vȣV t'](-'PN#tn2# N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@:Z'DEEEN Fy[j@ON B%; b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N vӭ1G{ݜRe{sOvRr~g@PWq mE% 6.qУTDW,+Yj+LuLWHW}-c' m[xT\rCτgo |-0efžY}&/gW(,3~}lA\ye@ }ۭu0&U jhF[ MSJc0e+kQO_B#ҕ5ڙ`D5tEp6(5u"[+LWCW8}EtjCW@:]J陮^֤$'@M+B-U+bP ]Z; c褭l~,4u"PDW6ȩ󡫸SL ѼF:tu`߆AiPӢx]E=e*+VTCW7ZqtE(c:B v]Ni{ml)Ec9 E:R'{@ ▦?/!gva].Ȩ76 ?k10sLP[䍏/?yZ~usI4Kk*l9Ӭ./gKWjby(7mBYρ&Θ>/' [y~)\}F/g_//ӵf]ggJ #}һ|IuMjޚ&Z[" 7@msTyZYyƼpО&P_>?gYUl{Ɯ{Wn&틡#~"Mߘ$b.bߨΗQ6[ca?pfd%{hƴ5}CW}N݇`7ÙM1]AT7F:hiɴR<#xlM+8ԳJpeihS_Z J.-ho6~XtEp 5u"R3]!]a{++uEp}5n6j9u"1ҕ)Q]Wxp;]UL(`:Brt跭تzAkY"^LeLWGHW>X[R;GY]wЪ0yuF{#5m:`#tLpU50u"1UZ `+gV4s+#z#N1x׸QA: ok? wovuJ&v6LWOz%z Z atS+^LNՁA3VK"@E%@=(IUXMk4ՓWqb8_pڪ &UIh:]`UהKW]nTNWrb^>Ȋ $bІɫ+B3]!]yIǖTDWlU=Q ]ZNWc 5-]\S"L2^:FJyi++bj֮f]=^ bpǴx~1ΚA X҇_k0C'Y/]t/ӵ2Mɩ%ηٟ; v<~>y\]/@Ug/b8У]['UW{DEe,^Z /bvp|ww8ڭ|o_ξ[O?A.~_ܽNؾd^^Ax*WM_j@h%fWpҜcv8,os]/WE W]>ތTazS2Ko}vMeI_sV{#H++!I7y/!8=/loI|H \ˊ~Vhya $}>k`tPF{Twr5OVV. CY`>IC:[]V7?@OI;yO)}Z6x2Nl<\J?u襴zXqD=>4w>ʄ-Wܫ民0[nhM ڤ\\\=6"3J%ϥ!͎;8^N54|уIW,8J[^v*v(sSV;sm >了na#ՠ,G S]o}+gnӦφ9;Um/R8?|\Fzeԧ۫W4N:6j'ƫ7|d=7w rkMuGK_v_֯G}ǨMrvuL}:d `F;6&=xg>Xzu3" uU T' -yxXԞއ} z2RCo:!4mOpߗ֍rj׷޼f[m.}ju6ﱢѻg֭|q& ' 7 Xc{p:SfO {]Jysj|*fQ[>9zxfVsv}̼v1 7g)h$M?A[};k&}Q&‹27cvI{u5 "ڐ:HejsWW{ѬA.8RHgA`Ӓv[(>f:YQ^#U)L]G#0Tz zaiE&r7xP߮ =5G=Ъ7E5jvvVU9?i6}Y}z7#g1NFk//{O[PR}R5~YS(,e1A&(ndƏGk_pm.t*4QVΫ*UK@jMRX#Dښ)}.PpCnpEr\ֻq-a2z+cW:mW(b"sϿJUqNzaW$[.Bi"Zf\WQ_W(>ɵ Hq*PW=ĕ U>{X$ֲl Z-L"2W'$J|RI\pEjwJ"!O\閦;%Xˮkv@$p#p3jzW(ثW$\ZRL-*j/R \xUJ I7n 800bBvpDK7~Syg0eyi%^rIS5W,4d2geq ed?89@QCNKFS $6S $TzR>N-+=+ P.u\J+3z+9EW$I6"r*y\J2z+4pak+$\ZL*M~G\P0hWN8yq\&"^+T)θ!>3Nr]+/i4:HUqrJ0 >I$ײj8!|92-Mot(bqrMǵvUy1JR MLSM/C^1 vVFq S|\ʠ2z+$W(v&'7.BNqE*SKdȸ@: T3P I}MJo\ھZrQ>5e3&cTUFym#\`'lpEj+R]qZ{>`١Z+Jf\WV8 QbFW%T&&$>yLSY6̘ w$#\`N"MTIjSW{ "]>sv(WWV%]J}tB\\::Hejk%3:ӐStJo W$)."+Td#ZFBbB*6sW"6]WAi8 +ZW6$?DZ+@HBŃ(^t=w'Wuv|p×\\SM/% WRZ!\\Z:HW=ĕiq%6BJ:u\J2+P1[^=)=4ZX jc6Ln)X#1dċD `SEǵPnXs/~\T2w|6`}Q(+; v(sz+pѷ8olݘgu_T_>\_էE. *ߍ(@nzуN{tY. A 1d@Ak ?ί,B,|mvet[swq>HYҠF>rʀ~\ٝgEƏ5(g%vu΄֘8+٫Pv˫y9^ez4ሊGFR% U^N`7pt(R-c^x ?bZot`F>-OI7X#Q2C`^\\ƹ$i1\3:^;XWtekhVY(*>H.sI 'y]/Jh~kpl~)k!=`H4Yﮂ^U=>~}Jy@R$HkNEМqVH>Hnzߦjp/= 2HN; p&+RqC\ymK`+,\S2㪏W90 |RPlVZR4y0rp[*Jp䆎E]'S||SM/ HW(WIWVqE*d\W{*#\ v^1NTȸ#Hx0PP9' #PQ5|(rJڮ0Dl2CimtF!u:o&W؜Vi>g.6hVM{$(i\3)]C9\0|'K!Q(r7oqim4:#[՚`a4V zR7w%vfካlR%Yu^3?uͪ?] 5~WTWϓmO*do{m@ ]ۨ"]-l;ɴ9.3o<]:@4c[a{.p|&Go?Oj&W/Wv6;am]xx12?~7Ÿzs2_a0r¨rR|rRXA J]<85QztQ7c٬–]OFmԀݹQ(ʪX{`5{dFlh߳bu^^`=2j,]ۏ7ۭYAíXOG'\g\va3..jZ ܮhbqz[;4V486`~ i-ve ǂƻZ0ϥl^Xe;i,ڎrM IӬE p[&F5Qَx7r ޼va0BZ& PbTGO!///#1w3Dc1&(K| @.\C%'&ƏʩNLR*]+ir"1tcTUjZ|9;}H`QXozzQg9陀(AfoF)|H\ֺ.U$}X2|s}eMoՂC{˕nucyMeYerAFI/-Y9MxZV 8:G֊mu:(]#=s`톽]k9.`DaqAΏ2[dڹfIb}LX %p(aJ$]-ln?euQ-^sG=䰛M@eY,GQF2ov -{z-K/. X[ sԜdj$4'/eoFg/O>w=SU?>dS9<5~:'01zc?SG sR A %:I,bڀ /2Hma*!.EXj[ޛ}4P+N8}hcGʱc}l]qF&0`K!{e@Nme-*n1M 1:wA,x~杜Xֽ/-nGx֛e;ݺwr"@1djʎpEg';Acv /:= BIoB#1EF%90%Gdt۱q"'cWo"'2<ޢ܎e;Q}\v F;Z1K΅c9EĽ{uCl)bL(+ XSɠbLNaDawIƳۜq*ȦfǃDY`Bu9<*쳳} le 1܉?" q@[)x1U/et3ލe1YA rw;t"'/9ObZ$"VL/Z#.ȡ"rd)T5Uc1Ilj~o <`dT!|, FBrhdE#&M5:ՑBif(ٸl!,n!xk Gg/scY.ʠl $(TһۧBFB F_ex%o%kv\k.$p !9,7Qmg濳?~nDw\yׇԈ?jl3[<ݙVD*<w%B/z_bʧ{% >a Ʃ}e#Fe=r 7 iز>IQ)N7ńIލ{{Ė0{>+d<YI犏[ghIћQ]~˦FƲxNl*doZXVɱ. 7. K.?d^ZN޽/aqNG.e?=.q_qċu [vȰqs/MDzfN+&tK/.nxV =[)/OƲn. ׂvbJ(Tib7xeTJwUBAyf48&q#biI$$XhiRJb&Rl9--D վaԗزaC2&{;QvIS܉|= Ʊ#iie&ݛG0OQeO.fIP՛anU+:%[v/x.~ |)'Wa옐Q`DY5R.?lb%" QwX6qG}Vb^E&V;ꈌ͜"&8RpCL{͸QVTbᑲ ܨ,XvGղ"Y-Iن+tͰyGU+7xm5b/+ Iwht[_ tpS1L҇F(Jf`!K2̌iu ǖq77ԋ.X4Ln'J л,!0mS)GF:o^s [&!qei͞ ,7x1Ex8Y֦{ln(>;"1{} dvG&:DsقuNp>̰H2tbS !*kxv a/R=*P۬22{g*. MR!de,414BeQ5~,QH>,29%+|pJCJCzP}%LS uBA." ojP!"fb1>='{,e.Y 5Q6t2{P&ߗ*,;LGd-Gk,toK3(l)*JQ|#': $ҔeԆ$V*Y iJ:d L.^|cӖTTQq1F86NҌ$U8 c l9L1nN#" ?>:L¯(BY7.RNCA+UYPeYda(ڴv:n%{=jYc0o%p] _bIii^MVh۴zܿSuaX.%{VZD]R3y: IReFFOQU *X5O'0pB] ֢mx0S9]ZCqIGI "1lE7ݖ g_?tql]"x[H(szBӸιC kxܚ%{0NK;ݷrF_?_Kx:W]jâdOKmsN1h3+Wս!go\ 6=AC|L_Ψ]v׮Z9h$R3 _~wh-iܮYc}ܲMXƃ|X5HZ.h.n;YDl4^}sX*Rb^PB!y%~_~>jrwydxQ}Qqc){&^#͘MݣXZEz5Y.6Ǖmn_ ;;@M9Uw( QxDR0е{Iм  egА[UA€^?xj+!O!a`*my^yC9<:;K^(pDۑkĽPDB{*%9n?` Qwci Ih ~gq_Grh;*f(36]-&2| Vo߲[k_ ҰG*M ,qAלC`bHX ؊መs`M2ACϖ d$aS3PEv;n%X02#mg0 %TmMAG(qӨ78ITK mu( j!Yӵ>9(/l~Gہ5~xIy^3MAiWQdbӑ~e}V5k AYP 1Hp{ ZX@20t"ehڡo!z I$5mRmm>N!ƚݗč!bn91u0k1{bkwmH;/޷/@W''(_smɦLR%533,RE= F=5&6)q@2B W+J6`H|o+n9x|߅(rpų,JiGv2~+D$68e8/ /黑/QrMuObY_7@W~VE`_i–p[]dy=]oTQ))AfԼ&jbD5IE$ӇJ'O0Ԇ^ExڒZC)a@֬?MRYh1+(lh3m`M zZ'q+5(+]?G+SpGNsIiV=e7mVRGk}Vط[Z% $2ˎ=Y usE; Y R-oK!-o#o=`bjQL*Rs Uv4JXQNJ,B.vDy2foEԍg*uX,\VuiQKh踄r;Ы]{_tV\%rD\GQ)xL98E"~KEj.2͓*FÆL>{]v83]DCB}`*-Zf=׆Dyd*aV5YI(ٝq˭ l\q)쑓͓bjx+=\WnI;tIe4OK35 f9ڳX=[=`_Izx+4 o$̧ޔ퓝.FhQnզƐ2]D{9 e!&72ǛQV@{>ah@S6\xJme7nm#\)ҽi= 8ScI\9&x6__Kؚ*NVG C_oI6C̈́'*MMkZEݧK)0&WwhVkgzZXVe{gvgO"o[bѭ* 2d[QJ0iXUQ>#[[+KǦh)0A<3u Y~dN%!3X; ]U~sMqN޿4ƓT+D!{%櫗]\qf}u.vJ'5?WdifTx|c)ŷWX|/JυR.m-«dÍ/|ʻm8Vdhgƨl;?gRd`j eSaYcW&aXW~3Lsa3؏ì0BL<ڷر2'iA}as84?bWH1\p"e.vj.c"Izg9)N  ozHȎ0awN2+ KEhLRЧ\+~L;zè'Xwow2V^W"½AQZ3$22ꡘ(~?T[RJakc޳!9tWϨhhja{wbQaO.Ftt1yH~'8]@w=JY*i6[>gXإŗ6Ҹ"}`WN'Кa{_)VSŝfD_Azgo,H.1O{T<>n#V?cH{6ŜzY*j@{%< 6y{dE#ЌRQa2$FP #Fs"jɓt^i%9) z6䲡U˙*Ԝ''v=CVj{n<N{} #eCpW|=c%?#l?"JJS*^3J) 0=ŭ> A][,^kFjͥg 3ͽ{ca hhly+ q:JƃWDIDm#n_0#1*fMuth9whƋ*yD83}\n1N08?tRt@&-f'RQqNE%P1r}||G[Vw1HﰄQ-[[…FǪ9KٞVDh pa% ~3?e1gRk'fa6w2A"nYuF=)歫wbR*)w?n DQ1@ycǩ1/n܈8# Dr`AEȖ^FK i]cX3tC @ި 4z;0  l+d~!;'.c2}|J+b^2.X)UAa j$lB␃[6޶C][J`MPBu=#%ό$ a\{eL0i`]6A!RW;z4Z~4к݃ T<.i8anVlCѝ*^qHsECo!M n7ʪDBIV*tr %hNJ ;3Ѷ+CPȒGgpX"58Pr(̣5?hU/YHc9Jf.+A1̲5`؍W:g}Љ7኉@Xd?f.ecjLel)ͮ'Lax[ɺ#^<ͺ2i{+p" n&4 G2l )X##,"U|\ĉ@tWߛ Uˎ5 !X]QvU4O-Jl@c>uGhX5ՇCjK; Vjj̇>ccq}'K~%k W*+Ϫ 6xS :>ac[0k&:}EV:ClG4ˬy_$Zz͊سMD%H9Wm,b{hrj[{O}FIҍ$sߨRȊ}xcÅ &ǀ58F<חUI#pDx7" .4 1 زrk`M\2}Lb8up/GAMQB(F͵o۾X.~~?. c;l;VO܉>(@LΟ\Ap%lb:hs=MR.JS4p,*OcU͐ϖÌjkg~e٫xjd)%GS:]5wFxIeoϤОX|sd4w>;W:M9TƶA< Mu,~Q_ 9OLh;w ~gD5 bX*޼aP5ǭ\,Jro=Zu`kz6)p~K;~mxl>jA!+8JR$9(g+)Zq9iW:XRP,jsʟ+%}zN֣n4?1tyuyV4rWک="t"4>8- {VKD5tMk*V:_{0MZ8lZc JdQ,r^8C0 w _~@)-SBN]m+omػZۚ)ҚR?d|B18@\ 8Z3^oReÝ[8r /?V] P6.gSNr9qK=d+S`kSb?l=9aH.Q+U#aU|LP!}*8ֶ.)NiߴD>ƒG/qf:m0 ыvGp9VX>Cs{Z`OғLz(&? {V͗.[(;\2cRbwRPE h{2)ʰPM+~崣s52'pj>-~9fץ$NMPCyô鄡4U 42BL"nb)Q & _cwÂX}e͹ޔd/.XƈAe*ʽP M 3aBW#W9 'YLӌy& 4>fi~ڐj%UQIx,]4ZGS0I#e)),RɛW%H[&VaBf&H7~@Sd.QlqF1΍s('$Ήs@cwvhu pQ2|9.>\̌2qj2Zb,E 4v٬q'"|\Zx+D<\RUaZq h|-ǵJ"),ufn\ˎ,ب^- :6ŊF Nk߃J4)46׷y:Ez2P!Ʒ6-좃{m)--"HB)QfƶG11]1HJPx^mJjAE_mKES@@}/Um|ua$YgL~VIKC(z O+w췢%g\H܃C嶭JF2]dլ?8/O O_zr 4`[THYh "d|sE4P \ %;C#ipO6̫3,h2/p>%E.$L`Ūpb8_u1Boa:ڤe־Č]>Uٕ#.܃ǡQ#?o]vAO- pPo^}5cK!Zǚ,!VvٞZdY dbvY5{+7~<V}`O[Y;mp v9ݡ`!ӽHZ1 }->tqIMP@syI; d3c3Xy8//&oc320Ҥ6mc@n{' Kg]ks@@152`!Ħ{.kA8̂? t)&uJ*V,lb|w]Ȏ/CJd),aDaQJbװ<WU"h_ärW>LxMX&M i) aX>k؎0a`WK`0kK!{a_{RWDw{K0! )u7 ;h0(';!xié- dם(û3uph@zIa7d⿥㿂Z>Tݏ67")Q$d۹K-r'kBL^%u޿}H78XZ2ĚOw-bA™9]  OFfMƾzSM v'痣q~83ke۬%>z2z}xNnc% iF=\GNzs&5`L崨^?o%'ks3Q?PFW&K Mxrfp(dN| s~LE=Q{&P^DuD-FRbo}ٗ/téҽ})Qrz]:+SS񄂵`Vv;$%K9Fya]^]=?~:iЀQm zqJN[R;JwHPE飖F'lVƎȒs<RCPNkhw""з߀A*¯ԴCp%N$Yj% ?A@Yn0JVSמ MXwۭ7 5PF7-a8F0x$U;ǷAOJi+@a>U9`ehFw}/alߞneZg 0e5C4Bc>k\(ST)c\ƒ530c#}>k`< 3 *{;tl0f6`4.C>Nh}pKkED\ԍ© LD-sEl}O6dܟuڙ7%CvhF{Ric? U$P\y,ۍ+064TOYVêIQ_3Qk,!J@_G*0~xc%ۿ- (L݅=ՂCa3)QX:ch Sٔ$NRbU 6Y0%#Pw?\+]4]hI<UKipF tI p G\yU 7t^yxl&hXBjí%10KYɠ'8 /exi4añJ0!l癮)`+7W LÿٳXMאC1YG#%m˗p3 $Xwkض-9cԲ݉&oMBŗOEՅh(&!4%6:Hd c3/vf&UHL Wh}G&ƫ=UAɕf=UwЌؽ8$C9(0K8::zM繠Fe(t%I19h}ޠ"ҙS E1! x^bB6oPl R?"T*1U1 Xwn:G'hptf߭ dznvv<z¿u/9K ӏٝtkwj~-0 vnߤ}uz<+0ioASŔil v`)y)X=h(.X&Pr\{L֦VFS# EhʓD8h (+h Sӽ{ق^ՏK8QvRSD uV 1pmi*چDbۚ<8 =Yݦht^牫@'#@\"L *23??ΓRx>C g6zt55oؾ ^XGU;=-Q!DO ?#8rB\$2FP۫Vk$M^|a4 SB ېG$OS 5B J?2T7&/Ȇ,%-ֽgJAOtbjdkF`X2&N}#^Feɾj-UՕuu{Q#WzIkA&U: ģ'UD&X4~&v$wJl&e,ff|-| -T!e H͓6wl^ЦmrY OuŭPwrXa8hf3[ƭx['[͘p3[dR,+ㆀA*:I*BntL X[s F8c(Xu=0mi}+wn2Fj-c{.ȸ)^NڈBP@Sf + G!p|4)okYkXW& ;Beo#Ikukh'!W>]Ii-~ ]h7;(,/[§.9-Jv&vA|7DVPQƓ/%̎T㢁&jq+q\R%0 $:H6,6߷ʾY, 1E%K%o'L!"v T.s {gP%o;߼}lSp|jI}dU.jSv)~\Hp;χ[Tʲfr-ǽ,;9=YZOUⲄgU3eV\(Ś.W_"Ϩ1pˍhä݄ip &Wg@iTYIZ{z6 V/d?:[6X,٪lh98mo9'g'!( A@iJCP}w'_V-"~HR|o>pk~ֹk7唡, !+ˑq"^`,o D:&RL`eBtJWF3",^3\˲6P%;C{S<&cԛ쭃TʨiBpE5%޵+׿"E/a#Q\K[Ǘ{i+lGIޡd[ds{K H7əp 9\#q{3^@i `<% ?Q]%ߡl.'>Q2]NKE\њl3K"-L| 2[V)Z"[H(4IĦ0%A= 2jt)%QYX\)m[BZ*cgbJ8ȭ6RhnP"We.>)Sl16HFǭ8 f>`6 ~j5'X-knqe=PfCMƽrUhˆ̿s sE*RygMr-F(*#8 "5ANHMfӀGe ! ٠rL2X%fVy R ^s~mwت$1Q8YDt`@ V`1:j ^I$ES<0'k6w"M0NMyK"WL1׏Ɓ#K0 :W"&,8hoۤ1,p)+GNI#*(t6:!LkXRG׏c*\(W,G͊Ede,FH^q GÂf`F(e2CFb\'VqtzXa!=\>FƩȡa@h}sGC4Ł'D$zĉr\OK;'XIP_PmmHbڗRk_(^hG-K @Ԓ]2*bvSPF`JYNTN#,ZEG9zvH'p|$DqiT| 8 f%e1|s>τT3=#%ڑCy`}<.tWo4\GPiY 0RIDED~8GLV1QӦ:+#4P]?R75R-..*e?su=ؑ1aGu;ĈTb =$N3u])eڪ#DYpKHim ζ6f)>+07P{>V 'w*Y{0TgQsEˏO}0Ӿ;ިr{O=}zZcuV!5_]W%[ʿGո6]/xrG75ߟj<(qS܍%nWkd}vZvwV}.kD]g>n|myaTĚv, Fݡꎿb'UPFoQg?QQ {]m)G7J{@- mv9:x۪ܪ f{;{M`?+.g[٣o}X6tqMY Ôo@ܻL`s ,wn0Oڳ&~N բk]gMӚijn[M,Fq< \OidsqT1%{BUO繷/Lz9};!E2X/7NZZ#Z_kU1$ {R8iEǥM̕SN&;27*ziT][Y^ 41~K㊭E g׮>^]z?-`ML9H`ƿ;C)\z:71G~4|^^LfD? .fj2}ːIP&e /"o}x94eɷn#j:.C[ Oĸ ^CƷ OMYFhGƉl8[ϓL9%9[\g4 kƇ9q`no##0mg_5vVci=XoU} $mi{Jk)&{m ^tP]ҊR׶尵Pb ?M-~I".Ěq]DpBg4E0JJ0h8$9c0h"`0=`em`F_.iR@ڌP~8*g'mD 0QVZ [?A&L|,^/Q"px}+l+7L{2.^NˡL)q44"@ĂGpaZfhKYף%騁]Ji`yۚX0R5ds -',QF1 HfKVE";yNBlf XXquZW bZ+jTʅh>Ԇj! J+5Hˆ I\~8崾w2d!z|h[a+Q!+)8Rxģkf6aJQު2^c!t8 |kcFBJ BKwt$RT\A,gu6Ez)aP"4&#Dh+b nu+[j-( n7Wx?J&+q&=R; @H 3`ڡ#;9`9s5].6R rjx^x;+ s(eޣ.]za}Lp)BEKZ>ݭut^[rt8#2 9tV:@ǣgO _g +J0N(l2i輤tSS0`9"ͪ{Mq|=pr7H>jʰ *FתEJQIy.?^7 ڨӵ:X7N f6m@]bL%ZKk*9smZ H6>~MޒX~Gp֔hYqYmWҗģMXjj>ش#rqGlj!&<8dsaN G 6Ӛ.GW6LQ6BwfXIC5 R,ChrʕeŎA%1v`썹40Y?ie)L@h+i˭WiR!jNBE@2ƘO(2a:?xvvffAcZT_-e6$Lа"iw& "T3e=wR|M$çKxe{Iu%mשK7ߴmK(}Ϣ-?&lz]e"?B=K6,EzZ{0߉ ז6sӺ׷9TaÃC%7,1xNYS@v8w2T؎TEA* ލ 7[zG,gL7Ghh6 I^hHUh'5ϻ"pn~lo~+vx}b>YY8G$i=' |hc_-i}N֦H7<)v'叼7&91s+[E [$9ٸ_Q6޲y_C7n <(E;k>I9vբMU-8Ƣv q/Z|z+ F_#sq\]1.>,Ӱ.cN0G+G\5.lqnr=d1-:W'l(f^TMAZ .*"i}PaL,h:}{CPdF%'`('^)(+9`m4Ee/xzig)kd̽܎GpK=0ý?^ˏg ^> s8y ͐d^̃>$R-{I(hwI0Z m+D ՒhIuF *I R) V5,yLN!&wZ9T.AQsZDcVZ"0&T^]| : Jp:^% JmɆ2v0F5y a.~Ib*ZbW@6Y %ܢAn-i[޹ /].7䭵#Ar$<%Dy2曝IϟqKo Spu'i5|i{4%g w'F3y_&+畕ЬbtnىfcN=ݱOGO4hv9u%@/h='OBٸB¦*.]U_d}1z1`nLqaVt_z_08H85ZGU7[<$G 0kSm;gv-RvÐaR IʹJ?(1dJٹVJ*=)Qx?p^pΥF 9r 3e"ehX9_`¹4a[V8z8m љ]Hk>5EeٓRr,9G~<.^źnjSlcŖA'δ=:zQ3~}TZHibjZɻpS21D'|>Cz c=.WYfLLkjm1XV8Ki5QEWxwƇOُ.<-#v ;=1oyZPM}ZV,`59n EcuK^jjsGc#H[3՜09gIO;)}ܾM"ၽ-K`'"v ʼ ;72mgc'73 E d!Wz2%(џɍVbDU%t5 K@+-~MJ'!9}ԳANCվXLRJ>xcU)LSSJI9ѡGQ- hf.H!bdC6CarSCu [et߁y͹޽Y}|5.tū.ʗ '5> `UT,ߟ]Kȇ^?C";Z!)܁^š5cļ [4?nSΛ@,ͣ ^QԎJűc1gilXV{,.T?_>;/99yVՈO'/Agp{ ᐥ%oC?0$A^S}ÐʋÞgUwU:G=}#ٔV- <O><2ck*Zϒv:L1}'\u6SWv^Wh.>漑 )RF\T\qVtJI4V@('uB#!HCl!kPw%BOA^h)$K go,G5xWCXCSLy~N@~9YYyCEݻSWEb; 1XmT&:P;Gue3VlH=&$5F8M\%e?-׎A!pIt1E_>M Q,rm$k#܃7h Epև>h@Ѓ T is1i_O_R!C^\guWF^_3)%鈵$gƙs<FMܩSS=`'O2h 0x?k Y㈝HKyUڔVɴ[èϜjjLCaiXb(V::ǭzvLa殻&A"s5$U kask7sZHo %52%],2s4`nCE4BZ0wN)]>88cgQUXrW:ETghB!WAy CN):o뵅uRqS5=uHU{AJR0?սd5BvDXo w!6Fiܯ:MANPT@XwkRJ=pas ESɱ E_`@굉 SPg%:/c8㠶C}&G.93S:*ڬ%AT,&f$yOu2eḰzV_Ud]:;FR^-qX#[&Hf_ka5n(D#G$1ʖwҁrn5>WwgGDJ̲p@gbs0fm 6̵s$ڊ[8~O&z״c ckLw a3^\mѹX~?Wp*q5kqm̀ϝ8{1e68+ծU-9S\yIl,p_z.R-GGm^U^țQ}o"g$ن'|:<򙼫B3o6s #.z5`-il9?mqsBg Jݱj8aYOJ_B~2zzr膕~xʢ~+ Zt|ݗ/4H@Qh.PVTp7$Ćyh[+~f\:Agg9~Bg,!}mx꨽mZ-zZM0 8SAsj1W\sO2V>bֿ..PA"I̿]Y5l籕cQ#b)>jO0\O«˭otEԛ#H9Ro9R7H;)!Wud2WTBLA4=υ A1ou6?/U={gRVՠ-~^z8DPEY>-;]>s)Fyb?~G֌g:X$s6'dIX0hhA%%!oQ)G^d8j5u$*)Tx9c6kTxcIQ`lybˬNṷuꙻHA9kZ6+W\sO2JT|jAݦf{X [o֍qϑQV7Y7ҵ6 8d MD@|e֠7e7 hg [X^]r1ؽ=8vPG%O;X]6[@}5v"{z`+ iR.˘:H;v_/:Z4Hw,K̢"x^{^+횺{V.xԚ"GG ѥniZhxu`' G'/e^Ø>̷e|;[} Ba}8 mMsʣc2KaճR\.e3b"w :J]^dG95:#gP슐k)dEEWkoro0ɷ $`.&ҽsZS{r{lID7sq|E*6l>:J6ٻ6r$Wgǎlj٧ NmJE_ ĢDH-U;,`Hdȣ4]RbEabjj _ "z54I@/]UwUEkq9rJr*]XNrx)Td鄸I̦ +R)d2(3j ͤ4<4T@m1ۘ㘓 D:6x~KǐD7nI}2_*;,:@=[iJyP4$rf'kw ނftѳ%t.f_O?U16e=e$TKY@WGkp)'h9b Zj_[\ orX]GkYk42dl@V巼X?ŲluJ&*x$ٔV9^,!\6:'qD5PFxNZla&VXQc_>yڽN!+b[~<2bBƩ\z?GVĉϐU,ڮqm =mD!ޘKJd&!QXZsY9k[/soz*!aE3TjԂ-j{"ƭ2Y 9dYgoyV!GJ pԑsDE"0]BPc^%T/gH<"V G# L.^%/;\{RQA-wP]6ʻ$0gDm+fg筯x#Ê.J|cՍz{d<QC%7#}_1ਚY[ @|m.t='OcT;Ќ;.6|.$XWO<7{Ek@uBݵLe!AFQ$jJDۜ#q#_(`>(޻CU!'&J"$'dz@!Id ;c#, *O( WiLL_w7s@Qr\ gK#pAp):EXmGL(PMe#:Yx)cw"KLqKz.Y7ߐ~74Vů.~X,.߽3go޻wj->~cY7s>yqCǛ} ]9 _ZoIl4 }+r2wlsC)vnG{Ͼuu&q;<k)0mKDϨ,b ci!>rӴY}wj\Sm4CzDZL 68b,aޛ=v5\^g.˚@DbkNܴ/ȢOu?F,ٮd,S߉gW%223]#{= s;āV j< $,YmڂMė{v|Ie* U0X놗ׁLΤ!L* lmea? A?fgeW`6.+ uAXHb͖eA?rCw;Xxi]M;B:`ˆ^̻k oCfޤN`2[Gib#+( 6 Ѯ2z3hWHOv%3%cڕik޺z~,ォe F?t4Huz,B+ER../GXנ>[0>2W(շ?OGMj~OpJppkZf-EݣzĵVXf]΢w0տRE0믾ibG|=\eu&o>!oٌ#7dB7q[}7{"ei; 0+~=f̤"`ﳵZ#vȋQs ^ig|Gȁg /oy Sɑc`HPEvQ`@vl&GH魤! 6us'Nk<$N4#fC WkD9pLS썊Tz/zzċQ#4Cސq,P-Y)FĪ9&+*~9W3XV"oa4 v㬰Bgg#XKZ;TL0 fBou8$SR;3-!ןI+fBo O:v|K'(^.ڂysn.)Df%"fMۄSoq-|ӕ]fmQۿ} mmC/ӂP%2E}NWABף)գ} d9ۚ@m9AB tK4€ntsoi=Ib 'Qsf5Zˤk}x&G7ػ3&g혽 nD4 :`urFc#2.y&Ü$St=I7C^ ' h}v[BĂ^\N;w$w֭<_jҊI%NyYT{gG޳̡l2wmCޥջW&j@Z&, @"ڶkhcwwiWKe7DRyI4z9oz7"JK.m7᭟qژ0UaA%mGpaA[I"vu*&*gO1M.~~[62`.аz6mr6-g9d9Dkz 9w߉pSc8$´dz;Rk1t>3As/)n$WJ <-}gv%K Uj6nLmmhlkC*ŵR[.f :ł1d! jZr1d[!\4ji:#/i|ELBT,5W1|N \<$[%D!((Q(kHJ,<1`ze+]d3<m佶jYL"v{ɜ'ZSZdk-}A1 xC 5+gI V~ɕ,`Kbd[?IsTw0N{nyY1 x]轐䱞<֏J>#<{+}.**к{WBMiыbK5m{1g hR Xsbe+HKRsIHl #7N<6Hp*+z?JQ(vww n#3{aqYS]\zW+%f T?a)ۡ{K y9}ϦĆXvg%Ky K 2GCmZu>[ʏY@T/y77V vR `5H5F-(߲Q=[U]bh*W9 B2Q FWdCK=za-ihe@g$g:E>^kT`޵ƕbeJ)g00tM2q.;TlWʧ*qڗυ(P,V[z2UHh0IJ]BȲW4XWJJsSk_cN&P5pĦڲ  VS6 TEUбZH-{ #k*&J)d[@ȱ+^FK*ZKM7QUAVVUnC;q>)r1@sTej"s`h3 Rdtp%+gs VʻD%f’dLڷh[Lg:Hn:wY\ y:G8T PHz#ov|Ó߉)O!}ϲw%fxoCw[c}d:kd2 UVӖn9qM}g!p`9i`~4VJѧ ,.N$r{6ԝkSPKT$Ln%-i|4#XBks"G)9KKCE\c Y~$!:lqNb(Cu%i;~|_TՄ֚A?NOTu튳yDok;Zg(s)DV [o.[*aN筩7 E?z|l)|3zU^]K#Wx龲LWPWZ$6?sC7C7W?}#*&xN NTgv@ٗ/Vә';v??huf4/RSِZ|< BV"3`Τ).7!$! 峳vxl<՞v I?t+OϠ I3Gi G)KHz IO7{Ά[UTnn}1pkU 86up9KAN\So$fT WB談sJvDd -} % 0"`נs1Nix H8$ =#wʯآYji%Izdd*IMkj 1wSQ]pG^(cF F0 xIR{onbaUcDs^-2A7G 2d2TK2"2l8k+ʭ 'C/ڿ^qp>Vs{ Qg[~$fAK= g̙T+;Kł3e,Ve PJ1wVbPژ2[̶  x^{~=u9iٍDWH'&|SA{~8ywqȈX{6_>zTc bjhf_MHSa?!߶bK[bRZ'R}}WlBHZOV" bV#Ƕҹ{kN%]Q眝yqedyBvr>f,zj #O7+6NvRU McV1ޥ@m}/8`;/1U׋㤿mI*`'1VYh{K9(NVa&vj^.lF%-".j+j蓋=&h2P%e']ժXȳh3 +r{/UTLT ㅚWzL>S*UZ%ѭi~27qۮ8._%~⧟*sO/Wgޤ߷A> z~ߔ?ݙ?ߧoSo^{Ƚ~1][;bQ{kbȓډQF :xo ^S\`. w8yMѽA-J𕌿NcQ2L7ҝj!l%o-"*U.EW -Ŗȣ1N`WÜxfdεΝ+Uk\P#:EGԚR?ƃ)9tfyQnS;3huڷtop:p;5 ׿ѻMmQs~aW0MT{X^o@1 p/[[Ӓ= Esl SQ@RHd˵ĶFHN\) }Fh]-sj*]IK7R P?tĪ1䒪U{ E> 1@ "ϒ;v7vgoTfr6i36Ő'cUé~[N!H:+ dxZQӒ@HY) , [^N'| %zʾ9ёoČ>pxzčU_G5nlc׿efӢl.gajN>zjvR͗_.rh7cdFUjQv}$\}1b .^x[d;t~&qː]֓4b}oSm~xrwIpa &m '/زiM?} on/di`wS0Cߥ7R(B~g\m)'2EɹAEWc$N+=k=b)ܰ1D%\֑Ç>ZZZ؄Y{usOC/{Vjr/2y%..QJ ri?#a9ݧƆa>i]7δ>-~C8MߥU/m bHUԠKd($-9hTѮݩ[̘Bޞ D̎ נ!OFa 2.SkP-(B Wߦ_8.SG:j"S˩ݦo4*՞y#1] y20@qed], .yF "WžJ5{M{18$7Z{8 zV4\3od7/1ѫY \V;2P.s9≀C;;t b$ !#L7F~"WxTPdoxUE|p&ܝ$QM~B?ޛRd?ңq!=[;T:mOe&bL7E󽑞-В 0Ƀꇜ.pucmk5\9 ֽy:<h5ϫI8}~zߔ|v~`GF{1I8TwN(|0> S8jo\Up.F\gr߶ 6~Tz8oV62UhVűSGmu=W3}yU_7uo='GG\YNn@V&`o=?|td8?|]sm=VDž%j0ѦS\A-$Aҽ͵èqE-*P- of97wfӗ˿J/oV߯veç0qԓIz}uyħr+b>>cEዢvn&#*!U;a4Bdװ n=UmսuyfܛZV~#]p p|G p. Zmh{ݐۮ թ9n^#op揃1 5ߴU5RȱKOٷ`Yj74W8z6r =g/æ7{qr/TTM¸vr NT=X/_$|:K@;v?_}ctZ^쒆kL_DeWvGtDŽ==r*5n~ =ZѴ}۬@E Gvtԙͣb [N&V@v<~@-3U $-nyd#;܏E{ "Z,::IݛU;(( #[2;LnYko1zCKt\LĜYjһuE8fs !BѦ쮅zmB<5#ԌLP32Af&>B;9G>Hwu _WWDǰ"sRs5IS6:ګ 2v"5eI C脤JEz!_) b)gV͘<{o#r_x3FӞw^lp_uojļoë-ސ)i`b鎜N˱^ZRڤƔ(Q) ;qη;dAuY B\z4Kk<>a٪Ys^wWW ^]w*vM )A=$gz9_!wF6;ژ̌A)ӭC݆1}#D"H1m7ڒ̈Ȉ/"8E[c@?Llu U jq-e~;;([xԕ"=DlIH8@a*PEQfjo-=L8sy銈dvgV 8 aIQ0TKaPGFFuV0:&<&̈́F܇T{Z40L/9"&8HԄ0 !P "zc4 '>:jF)G+xU@Š%9D;ٔtp դӮ]ٰ6}8vѴvi9P<[If VÈ#Qu'O^n%Ya~ \_YՍ&D,А<)%FsH !L1DYGO\+g)l HKЮ&〔:Ъ@ bp,I:M!lӯaʰq9ʂ Q-䫱*t-2x1y J:j8ĦS&S.L[qL\{2 49Wz7X2<턁=?jt>m" Bǖg24Qʞ͚]TB !'45uשrw*wשrw]>Z,z4|R$_fvPQb̭mJ{E ^mӋLqx/kF\}nb~EvAZM(C"+[@i``)\\ Vx{G@ʊi"lB{~bdbB[$K(H9D īy@2$wc18F{EAsNV= V{<x̀aL$wvh+:Uݺu~cmOKm켱)V5\i+)ԅlliIP/q&If-R-VJr<3-Cr*-nRh&)!3%xSuǘyfR1#P g)i8hMK%wW}(*rk* ١E/گً-9J8c2VoRݞC;dUm:O*Ю@媏^~@vmdv՜F(X-,#TSvm+F5O'C6F@;+%u2JhWr'5@v^Ϋ9TDik]]vM{3ctm:]]Aʅlv>BvWc=cTv!+r![2f(n硝԰=<$'[uBY]vܨ% DY6_zTե =v"q3 $jgjտNNFd2}^9d8mdYZ;sV/OKԲī#T0Qr񙨡xZ#\SIo%BwldFTs"QzMt^꣄e1F!ڮGӼ=Q}̯sɘ'ף4>s>b_ 4\'+{;/q LFUJ1cq+c[|6Lo.<[k[_HR{'lkB\ Ïf =чbs`UnPP͚)j>! |̉@O)Q~`Wn\HAps1BYoc=8!Zfn(* ~IzN8§2ӗTp~qzc IBʴs~0<;mv>X$=缊B^*U VY,cML _M~ѥĽ9;( B*x|+?D_"pʑim*nE^ڳQNV 9RQe/GX,ZYڳ3d(o;~|]7FJNvF˭ߚ"ٹ[u2?s%OmE0yb7ڇin(+s$1=pko?}uYtƫx}] _M74J+KsۮR-q<yrn/,(@e?0t4l_9J jRk㫳.L𸸳~yW;żFbaw?]VQӠ8gJm⨗̡>Z1@$缛xCfZr2HvK(bEujU47(a뷡ruMŗ!QOH/7}լPr2kM$Uj]9c!<0D]v׊ym=k]v5'݅Q+% v.阛/lɀm0hŔ_z, ,kb aN椘aN6bDH`iL)AcHJP^NZm }Q/ KᰵłCv(| yBiviA ޕk# \-G^Z'J-4-ys73`>Gyw[կ=O1ϽAzMB RđE';=5i6 &irǕE{n: j2B4y%^jU$1^9:TyteoInY0sԅvn%F^p4""F6s1oG.2{Lo7? ȋ=Y建'n:Nln215F5i=ycL5 UT녮)v훳hsf<?3aB/r8sۖ_$I'޽~߼iZ0as M0׌ C+DJťw. CAF+Ŝ7I@ UYGcu}QX`ߗ}4&+K3.C^Z9 a9 aLBhJ3 hIĤQ;MD@9_ZD:8qAFxK:Ks/_[%@DurLӫ.g59PC;CM5帋\m>BqZsSQDPz;CmvC-H%<\7+h ug oL./+oǪx u\n˕Tg"S0B%( db-$*MN pjED{ dL@`>< =y齽ĻQ)ӊF_9^!^j. X魢׈wyzk,} -w 7@ M""Zf :GhALӀf8l\֤49/"5RCd:`LZ7e\R"@Mjkt %_YʏEsڪXh{tӒ6fiZ0(6EOrrdœ6 x޹ht.hyX``>^NgG7 _NooXýϼB?3, ~h^jZOwb/?Ο-49 kgg92h2\m2/|V|BA_(գ_5Zo-nǚ#ju ?j>[8QZMcrc~ߏaD3btAb^nsrZDG+j) mԈx+7ux~-XǽGf^਍k MxE"r9H/jk/s k}~H4+@#NrV$ e yy= lMzJeŒ$ q֕ EkBg3Wh4Sc=$dJ "1h3En2F@6B]A:1o+йvKo^zs%moMy:{Һ仙~QTU3ꂝhl i[ V#m"6@_~/PۭvvFXB !qAex\C`3B Qt'%/xRi&;W$S E;}6"xW!,Y;JUim* Ho׺5ٛX rb9pP ;lj弭=$AhÕ1ýe?-Vie Y`6 \!9Ξv5هbR5ū=;5gTG*6 ߼@wC]חs";Ǧ 55Iʀ6gR}Wgoj Sh $) .)+ M kJodLF[-& #aڀBeu18Q1<+臢:PwH a<,(i Lpk| L8KNThtBG CKIrrl@Qr&bɓ ;213sAV^=CK/*o-J"AK-UZd:DE OyaxA0cPs|,Vo.n$>L*FՖGP1Jm׏p 7Vb@QE)&8b-Q}by4{ʟԲJry녑] ;ѕg'_8o8 .8PXZL4К7B6uAtrW7_rZ=#G܍zell<K@m- ^ om •cly&[ɖmi7ٲ{Gl:`3"Ĭ]t erAsQ)tfT|EUqj_ʂ.gٿT@dOguiR{?N71(o gЪYiqCBa=#{`X$pAN=IxS$a֒-o1̓n5qSb P8:vk,IY##AуULf-t Ud3>mcƧmvc]/cŸAEĽ^$&E9{ŭP>j/bԾ|I;D\'14AJhj"`$(单`DZ Mtgn+O&!%hy0lt_ /H$ٗ[G`U 5U'VE(r%Ww)$$V* uvz5pўU9a!2;.ڳ1{@Ik,! )8d2o);W1E)X7 A)y45z*H]MHg7<)3 Hh8ilAg{?\|~ߦVzdG-B4*_n}CHs\G7I_U˽DN-yv_xyrqⴱUj|tv9̔yf|vsQZ=[\ - =_ӚM h8W_ bY6MF) cX,>%9אXں9ž%(72C) !'u qbttRM rRB !`ԙO)ˢG5NiZv_pbKvxųS5l EaF<:ԖISi1@gi7 AUI/Eem!C/#kUr"YTjTk 5T]RĴN,9 dJ4k z:!0Y$}Ȗ{ohU`H:na+"` 7zIyov2]ʚ?/&C"9Ƽ=%FY KD0">2C v[_Ų}ߊy`%ۣDJItvc+8\~6nΝ̽*2g J[;jM=4aQ.rT~zyhv[k 93xM~vtaumNZf7#uZGdPwepì(8AV%*Kݗݗ'RnhtQ=#RLW67(Gգ5>3++׿^ ùy]}{>βע<}e9!B/d'zĴumWw۝gӥBB@IEKws~6g[=xr ,OY^CwTD'fZD25(\Ƃў)3'y>_ޠYNhkǘq}I&Ooo[ochb__/ E&h 35o+M4?y?&-l4`o#C ~MR8̡1F9k!5 Q$u),3'.w@UexFub@ NfUR_pLnM}~$|bQZ4P*ƿJ%Z1 YHLKX2AlY0LEMxE%5`r ^.{!e2" sNaR P,1}r`}-b5OKTM-Zyٯʠc 7aUx7"Z(,OQȘ!3)0m66>s:: 1sڲ r-ɠL#{v kѬQ#RaotiI{XL Jֲ1k_!,gȃ-\*`M!hc SuSOہHmѥ>-/k%ح^f+wƠRɴh2ʨ#*0 [>ɶNv[>hdHv J`h[+ȍg#\_Ɍ L4RH MY5dȌA`h\S lH84rR!NLT'e8( a^LZzGnv;iE%" 3  ;@[Ҥ"Mj4ըз+ ^npթzV>(!ء_ucvu?@q7xwBq&͇&hyL w6z.ݮ_t UbW֔^ȞV m[}ELS/4+{}|BK&^&F_ߘIUM]u~psuuwnn+z.Krwfkod\ w ʉRJ41Ku`'|pur%Kl݃g]*(t]*gT:lLcKo@q;5f:Q-F6 Bo^Bf輦P>Mge$BG+J#A9&6m#鿢-\x.{Ie7i].mdɧle_3 $gdGqU!Fi' } jȴeU| hU PިVb3̀QC;\k-.oY?wqk]\k9jjX?ZWirN;I<}(*l| Pa`Z;g>sV÷$jň nSy+oK^X{+-2D`₧ 5NVA9 dV|(jb @gbv.tDME Xu5ݎ_>c#T_VM"oFhcuw7.>ק<^:3ؓ? :oߝLOl>b%B?|$vpӓK{{}w|GJDqv֭ȓb=⨆nO4n.$}B~LC;%`I"QvdjAVEbK2#}x]u/UuNxfPɾDMyAxFPoϺZHlydQ'3q[\&zJK= պfd uYx|$m[iC><̋F&vt"ǑݰVn5/# 6MkJ-K:6_wlSӯypSyGSvNFyݺV&N\!7n(1PNH{|0ex]~xfqqŒX# r0p¤9:ɛ3tcy9@3<;'Ӛ  1rŒ*gShξԑ< Ԭ̊ţ+gblgǴmIfGJף :3NLNl}qyDTqJ"k`N ͒8+:G| l9 r[PJ-9d+ʨIIVg XZ njڡsxۇsD#y S[ #OO˚ |ɳ1"<$kc7H3Y;us&s2 ̶2*`ȸ;may SSӚĄTmW`"5邒Sri[ЧRlJλZ2/WZP T!D3GShDf)1WO˜nV~G(C)޹!G|yv$RYRR|! $ AP[FV`cx4뎤t^$Cl޵޲EB)U% J6RӊXK&j*U.s_UKPKUa*O:4n mhbwb(SEy{T%{|%![5j^qh*hFXɅG`#=(hv{PmȼH*yzh|8FR$=(!]*`oyvRU>єEJ`H*]mԼesЪFh%+W9oּxMQW/؃QP;9*_#wPmT Tik\^0Zx-{cڒUlzP4 ?\x24 kښhPYC%w 0x3(׼G?~A j7nP]+le(.Ve@I-Yt'jۑٝ"CXƎ>[i+av 7A;tOp̍C4MVpIhi{5WHlvtUl6vɚgxh|7tlϫXUj9D;*"*r=@{O?؃aK txitXՍqܦ'Y; ?b<ʹ|O\]:?/&Ynxc y;-vDN.Jo"N_(Gl3z u; "T1J+m/j%whQ͜"}A띕 +TL9啙ךv'Wvv{n;XYǰ\KPN -S8wěV%ZY3|d}@jrN@c׻ Czg-+5zX @gؽ1(I ie 0@cnyV@X!_ϧLeP !J:RbVliT >* R윖Turuz]RtS#jw5E<8pO\]DC6q!- @ZH gxc=ŸV`VVgas+$H@Gv`l{v螖 PznXFZ劸wW;b/6@v[p6%O1I!aw$6y*<݌ h歱<]587^nAjW20sM/C!ue>[?#ڦ23~N A|lG7JA/5In-=٭}'`z Y} $COg8LN!x{mJX3]080P/~ļ)YP Vjkfm˘یggc\]/VxXy:>(s"wEjG?Gy_&˚}1R%Zqs8,D4WJCtǵ޹ܾnY`1(ɺbE(*BŪ=>a _~E;rUʆAd *2ԑ*jEB65PEAXt-|e`'(=1J:؉ο[{ wǾNo>}辏 tsI׽!/>*>=ߜ_^TGljK17|"I(3ol ` z"MfӟsG@C`=e|E#)^ثH $LAf FKx(+N-Jwmt Ш#HNFt3Q@Z.)^-- 0 HCP.oha!g`p:|k1ߎ⾥,jdg3繢 ԍ&; 糧sv ' `$Fepf/YxCP|RaTw3c$z|6iŶl_ntW2EmeyOLN_ΘTdRJVhlr?jj"-^ah| S:6 p㤶To&YB`%63m%*ؑF(߄`DjLԚ4H#n%:2*&OX-Z>յg5Z FE`jțmF}m9%vStps^˧:~$1#^kۄۚ/u b "i߿_o_l~_ś7oٷwdzQŝ_wm}a|o[n"swpcܙm=9E]o*>DhN u'ٵnF1XR RT9jj1VRc1³ֵm֣ Jք:8@-lѶ9nm-HmpRa afO(l) %<\B^)MJl k喍il_Mț-`9J1H>(}F;ᶅ8yss=Sh7%:Ob8=z׺iwjN3r.)q1V;u+huBCq)S&v1u۴Xsn)wvsDRfbn1q,P]Ҥ[Gʾ= 7ᶛG?,{l7W\yS %O|:;[SEJJVF؟̜.15%sa^~FT0$=$~Wlmkt;-m7ӉӾ6\6RR}HY$ʺ:iK$.987 ߽Ӥ t쇺^Un1z8 "M!C)p+!cXv 4V1l.b_/?^Lъv4,ja'DJ6^+5,؞XP=}MBR8И!CIjKo%d%wL )*lH6Eej_3&( #S;C%M3A@Tҏw栮% f hRS0>\1XQ%V:`k{ҟ_P&M@jx%JͮA dK?x~>l0گ̪j ‡55`OVX)00Tʎ/{/^ *|;j  }HkI%ӹCHB $@[#H$jIS؄$\'oHJ x 5`nv)t/x 5PqrH\ ZE;H .Kg?%3MXpOŻ^Sl.78*O #N7#X86J\DJ< >FkPYRLGV x"ZסlNglNclN.D%1,DP$ [9~6ojDvk1(6rXwd|t|7-R(Icx{j8IOp#`ц,#! edüj>YOwE%<8y3C>Q=f Q BƇ ͍t?W)OÅ3q]/Ί_v\%['t :*d),|{h޲F3T"yZ@hX~оS깢I0-Ӵ̞]=W@ \s'{@=!U]§azMr!C[5^ O!r5DkQ>b Ԣ Gi/1 mGD"xMTkr2"HZ?܏Ȓ]#Ǯt/1eO>xr.t'ѣwNp DAwμpp^K,\CeuΩpY^vNW)-~r-~å ;*'IS򎭄QgQ|FU(w{ǀ-OW\(磞GXn5LP xgTv~{ q"b'P x;1i5hޞuJ{ x}hN!+CAZn2Wcy ݒBrDϠGp'vD݃-0uaZd%f$.}le;a<ń 8CkD@wT#Os,Y%Қ`>'ʲrF(y>ʛ:IQP Z#v^u9e3Fd7 dّ߆xƈ2ފ< О핫c'/?Nց(Ƕ6>#\s _ZHG-=$s^lbP\lW]H W9[67slJ}T|yhTz,8ΙC-@͹ ]w>gv9 j\H6jSq C:rvB+pv>0 * XOŦV*qۑa`7z_A3j2F!1 Mgd齗̀XɲA6^ᵰ{(+{ Xu:ŠLWv_Au<Ǜru wM(Gn'teSO\\V֖Bߗ Xo:-6)9(u6Sm\J̴e^.?7+|B*߳bϾyw/^2mj2,dr~o .ցՁyW=pρm8Ձ}]co`k#gW{5Vw&+B._0$Km \a vjs`YG '_clomloy RpId$1UFD(T5"qbH$Ji*Q+r)hѕ7ߙ)krJ_%.uqΏ^טn跥%/V ӋQv>fſ8++3zpIeat9.5GOh\'eUg|J@ Ds fB兴,+ɩ k$M[ J*A2g2D*Q.< lrk=cp;͔8:^Ei/YdI zӫFq1R,B⁗y 0+wQO!?w/ ?az6_*T|;需of1Z_IpAvPnnsAǚ6"O칞ąZzʠm 2_'XF `n\\;+yͦQ7%X)j33J^7[ir6K^46ҥe(a_>f2r¿OC0HvGHjUiTYn<[rUd}W0ynj z<@)2آ˯!,MM~Hc\ OOG `iN,p7ī&t蔇=a٬ʻxͰFٍBNhCXo0(H Pt(qYdzc&giFtuy JI {K)IEEX I]:`E<@J38єu$V($F2ؗ h4N4BK YfB s 1 QLG, MU0sU _Qa$.׽YvO8ñ?PK}@"DHv  d, $`" l;8c&ZT٬m&&\h=f8Q8[aIUHEcr:J$ ̥#-@Eg]gCK7n+^VEבkj~@AۀZޛ" {[(w~5@T*܀ٌvMCIvm ̦J!4uljy /0GUl& tLjhߥb6&+졊"a%1z\`CQl'F&bI&a$hofh" \<`uX;10yxX'M˳VO]SB60J6>`bʔ_iW!3[Bk/pĤ+7LlFmU(F6T(B@k%KugKo 0]yoӰʼ\,{,+/M'߽{=b~O#WChP.?M(o;$w@>TI[G}[,4; uFС/O HӏҩX8#[#+j+[dXzxt簋c#r׭̀;ȧY[,(Y"I! l1Ɇi6>j FqQ?rW6pNs`i*9gyʷopc +5P4l\g"`kp0yeVr-}hLi_w* ֶ? E#s?M?qE1z9,S[@})|OЦimVmلJ弐Nzܴ)NAWg)3p`X93TvX aTwˑ(]k=3ji'vr -k^lgVC%v<CރSe- 6)[-1l%UBn??̦4 mlW0}kH~0xp JZ#\.Sw}]1yG)&zchZ}pѾ?C"uD.-aq r[F87t:ɯ.y=a\LK/.?Kp_8lOL`yS='Ơ-aY13,7|B℡0"2ZHs@!& yVu<[ ^ \#Z$*A؂3CX1Uw{49"k<X$5BgX";λ HB>T=Dz/tW9c(@vsz1p6q0kg8lx WO. ˖Ni&M/پRiEƪ;8_Ë?[!M/ۄN"tJUJ+鴝N3Eg:V-)7+[/QQG0tmyD<57*M 7ɭr]ZY=<la>ʌOjdW!-_"9~QOpV!7#رw&s}R)rWnpC$4ܿR&U hy֢]EL;<@񐌲:7lߊ3·PYo.^#fX%0r˞HK(2*"F(AgqH!#ߏA.]}}fsWI4b{IDQ/Rͦ47cĪzwUʼnE\b|(N苂ℾpR*Sv8Z_jC+e_z9VMA$+fϼ\1|)fVi-6J#0NhV֗7ԓ 0t.s 7pz@CܞF{w_[.i&};;Z|oJ[?ť7?8FݛH u^ |g:"D Gd>tIE$2G)͸X:Ns'RLAsooJMsV9,/[O@&0,:;sf $T3U(Foϼ-ɭOŝ~(+cV.[XWwLk?%5n*rY{5 ꜯ$ռ2ifT+s"3x !fhS bi5WLN;71)焀@4 cQ9Kxb&`IZcޒ156\se{8ZLƯ 8-#F]"TRJUMr%XJh>)'3>BnTcE#xm`+'[^+қ.&rjnvхVD r gt~5">i> A_VtVS_nB@k1M?Ik"b,gk-Ot%6_fPT+D>آ/,*"!~+;[Ζ5,i!/zoFWq^[{hzrv<-n@͠?_ Z*z\VZ?P.S3\FZ/T\Tu<ћb]*jWK:`HwZK(wbv7XO}G٠a+y?Vc--|IZ`MP`x]V;dEpB;6trkjwM+ y{yØS q\p#un4pgb ]N)AAҠ 넾vqMw9Qcٛv FvC8WnB)v -MX'm[kÉBW݂mݺ'΢<%ךn[n4h:h.5E6npg :;_sGn4h:h.3ݟv EvC8ԹyBvf[( N;hKoi]([#Tu!OEA@ք/$x\W7EO2Px?w߾6|p(1To&'sw@m)eY/\˶{3|-&l~?yh~[b \9KIJC~/+5yp .>{f].e[d>ׯEaZHs\NnyFR{Ojn%}XuΠ 9cWyŧYĈY㧉M32+:# Iϼ\ҏ|ݤݤݤݤWMJgMxqջW?wtUY1"G _}ϳ;y79JHj0#NY"019b۔YMP:P*Nƥǹdyܙ+B N>H|wb="m)Tn+cQec D~)Lh;M9p~EGغar{o I 5ZGNya I0-lPPl'2+P&Rb)Z4hI U̼;&5Z2BA_AjrnU\hn&Na{C]s8 ~ٜSZۜHqP8EQ!a)m062M?BiX9ps;ۡ*R"Ot=$NQLůAcQng]!]v0(\[ ?T]R琮㐎4\ TeܣTF%B[$։qA3QIEc6ȴ1\e8-|gG}!Ę?ne3X8tJY4$Zi5YeYLJ(ݖ#|J#RcUB ȀI&b&0F4UjʌP6C\FXD>X6ER9p8SK`hD@bXI ?Xq⸠E{jes&@(ih"rn0 @e7:`FъHXa My 0 [zd ;?Zo*Y(ST<5H8=`kPӆPS\t*G.;6EgcP9e(I)xm0pke^H'9rAGn]")--W: IpB%"CQGg<|C&:M2-!3Z7s T{#!QCZ!i/sFX|%VϤq@"dCs5VQoi9jQҗXV~~0rP6#)*fyB sɸۼŷ/>!򫜍+w7\|| $\s0xCBg>TZ#l5fd&R3=*ĺ_p7̾w ΧJ%畸x(A@8uJx_.CqX˿3/+": ʦUd ;LL' Pƿ3n5:uZd@~1\!{<z>Ogl hª[]R~Vgq̒HfAJ{N@mo^]zNk9(Vե 3[+UpRG~ Zrw7U/.eDaJo&`!P 1a\f`i·Wp_ ɹ忟 wűpk^(ĥ A7m悰P:]n iqR:٧?|q>rըea;#DdjcFЂJ;u⏏dD(=u~[} H%y ș ! S)-QW!ـc?T[Y@`cE^ow!R.E }:H"~imoQy'oqDsR=u 0)f_(ψ~(ލ$HGq?8kDfOӉ;dG?9(EqʖS{C%U"(هi.[< 9qQkjڋDH4\!cUvyCnKLbPy;qg nC6= c'˼H)j9g}#:z/8NQ1Q< /w((Ddi&x.b f>@̀0L}e-T1{FJy-dG #&h\yS=Vtnj6) B)ёCs r>^TB,|ci +bw)Lz7C*2'ыPAr6 Xd.~m9{= MWΆc Τt¨uCFI[Y6XbAbh%h| (v0{gLiN;ف_ MdgtN82HHݦYƴRj3#mwOx@b'K[`"0Ç J-9U:Gy?rYM=g/P\K!{UT)ޡ(Y1*v\?n6Ȥ5bю[ ݙ"1GL4R]LՃ6 u}ڟU ")'mY4}a` I8/i+1߷AI$)m^#)Ѝ;HozHsUv$p=a M10fYXLCbN'tSN9 Oj3D F}{ 9pՠ۲F}yA (s[/Zd(h4:RGa'3ȝl N e#WsG#)LYdJBpwl9 ;0`PQP5X tv pH6Q'b evÓzo;H0 O@9sFi$9é1_؂\")݄Ş<{5sL j++}8b^< 8mApK%{C||Sܕlw u bUX,\BNy9!'gXo)d.c÷<"A'Bs8SeZJS2:F DPڏ)'p4*'NG_&`(]4ѥbL)b>V c%%u5QM}=]mKBaol2\t<^LtU}'] Rqtq9}k?BqSz܌юw/nCK?b8\Nx^'α2=/@?{1B֠bbK&}bj>`iYҥVCռ5Kv1IJvƦ69v ( k3 k`o_=IIe`ڷ9>,kQњ}gTT= gy2ot{"߱#(8 ôaU+(r3t?Cԑ~k٣v`z~.<,u;VR:jlUSMR`n=%ijOU$,z+Џa=t:޹Q˷"op}9μ7vetUzXp=:hT|&|߫< DۋK zȫ!{<_|N<л]aszon/^~~gdahdoϩ 7W"ty)Գ^+gXL=A㪏١o>1k{{Q6>Is&S> N]ԵpiB $ξ-ua;8d•DgUf vRdΠ{7,i~K8= &(5 g v_*[JN!rԯn/icG+##pY.R7$-'59ޚ'cG?v_γB$E3mz#q tȂsk;KD߰ VZb]Z],&YFezhٹ6k;w':u5CyL]IJ : }(;( hTgϧgLweMX]Ap2ꏧaqn ^( P@D!$Ia 6 ttiП>h$80Xp!JK0 4 D*0$I±T@umM} s1u\d>WdjBBsA|%Ɲ3ʐ8 oiC9qBDך˱-qڅM%twQEZ&8 3xʔ4VB 8S}Wt1!"Pygʡt2f=W.aJEa=:&po/[s%S ^}H3x1\ӎ_ p'k]Z@0q12y(A}j0QkٍR kr+"Өcr)NLn|0c=#:,χM S0Z>넌$cc#eLdȽn_R:g_V.\?krB]tB5wtۋ5I62]@M,]kI<^ ?UѭϼN[W34"U+XЇ$MuLJL$-]MQ6m"ZЌYUnD%6xIM|㒖o|v὘LvY}hdT^?7|`PwvnJd~iٱwfÉx#iTՃ ں7.m$B$$˝;NrmO')Kngy.HA;u} pg{KRȂun)H]>{S{o_u#ĶɴٞcXΤ 8O *yYbQbǺy5]6,-C;&߆+y'T؞E{Erd<\ඍBMh;WNΕ{PTX#]{W-O%^(MQ2_ +RSj{99oߣb+W4x?x[m ]QBzOsR<[!uˁҨz{ѿoS|d6N;x} /n/]w)=(UV-o-I7)=ޔJòu}iHeeQhY̼|X_^aКlZ(pTZw (TJ|FY޲T䉷LI'o^ U( Ll- Lod<)ltlY(ЁؘW[yhT˚K%)L1C!6~ni-C) OB}(Ha8:R\x$1KY$HR $H#sH ,!,ዀ`О`Q<􁯋uSRG&MNIJ9xgLB*NS61(^]]=R vo5U߭;;H%\[3gUK)KYdrMW7}/!i:%m1mHs.g%&[y|Zbt{Wbluг&*Έ(ңi*|85w39g#fێV_F\h|gMrᮠxӛYǟ3dz1KL''az 9M8Bp&>BԇTTיp>Z&o)H}jVsL8h͕$䘓#&S~\ /{ksǪ(ܽ2P%O`1>^Ր[ ۆ̹e^r.i7%;~㏫B87=X<ʏ׀{O$c-~њ~ؕ3;cڵW|)t̮YwC"!BD$IKaVL&VBv.Lq4pgwz0*9)2M3oOl9*.8%x wkՃq% !=-ƇS`̳ ;R`v(ڞLPgAQ+e;CǏ45b ="5c~F,pAj<<I;a #<,rt0ӈSСUJS:q&;%EDAn]6zMuJ̺7?!IHP+]U6$7#4ml6[~ :0rj{4m3I-3gg8ĜY+vAae9W#|hU7@@yw\g"5Ѧy &2f1LrPJ*%B&1C8!{U2~L8 $i*X2' Hir8d'u7RWFu$pIv#_4%-W)p("HItCXƘ04Rj+ATKKQ&K+9g3 D-rFP[3ف o|s׫ydL71#0@J-ϵ0˫{j̷hVZFMO0xBw䉔Tԛ89#_FR))8U%A4`m'8a@bb UMmM{{9])ӕE|&%Wm(lls5p47rA-aS=G rPI`㡖})GAL<>}!}_ȴr/EIJc IWW\)Fn>OV*5@!{zy/E1j>~~gdj¢wKbYL&~ `lScͷկ +g.TLAr}ٮl CjcsX@ -onY*c*&T G0QX8aZjH:Je$]Z]g B+2z__uqW0+gNC \VjN |+T[)]_JToQ? ^L]77-Gpu⏄(in~!wr]C|=mW$#!O~Hy]Y6+z#quD=ێ ;Æ ZIn{}*"Hjn"%e!֠]]`2\=~k?\( 7Ck($4`j_ ӱ ^/PDṧ/["'N'qtm0r(ģs*prWG `8q8^H)"VƱt(G0V))7΅wn(\cϽ]-Xxbˋs{ۆ>yՏ;ԗgǕk}}'߮ʬpTzm. 9(%*[C>ϐ@EJY@EC*w7[b$T/K3Dz4vUd* S1 Gf5HDp@ ?<}YLHP 8C35/v HpcZ^Setǻ(t˫{Gqw7r-#UR,/lg(8>S=pZIrTrԔa$ :{ݳw会}Tɯ3c=0ץffSNy^LrFv9wouNJk/): FH%OpʊqdW ZI  n ؒx9B2>q$ZAJ1FbIr9d^ӵ3A%q k!M8G@$txEuQ" (w@#!YB]UqV\|k*yp&$TGA/[5'$ 3DQt2U 燡bXI͸OI:8T__Z)d !䀲U `1yԔ@ƙiҼU0a>{ht#"qvBRg' 5a~F="a7[ig5C#}ޅ4GQM>D.g{0sv{(O,*XU?Ij72@ܘ LVT2)K#UUa0i#4Tb#&?iAR٭]hЛv|gW_u/n3cf[7a8x̜q1qZǼau&q *lu~!Z6:uBJJ  [ bà׃s 9_+*撦cC%kD~Ԭd訿]}zb|DLX^lκKF_KHRɔʁ .)"CJ=m\mTی:Eer`R4NzX\^$i$ Sdo(h"dG9b|;!;Xc.OjNA20`" W9DF3\y G) b\ HIA4R(:Ŝ63MD/ 3F\k1oǾ/<Ҝ{hBo-fJa>(yPZ **BHk]PQJAWr#ڒ \ʸP9PD .idi;km !K,/vz{ }ϝ(b]n]&˅ voɆ)@D(Hjs#PTdF#J2hK1PT$CђPS2*09J +jTmбH)Lq#j5Ҏ7$H$(WƛL1wf,8(e%xa-)8VfKha% ( EwP\d (-J0V*+2XSp@uMW @oFy?,i{/>SdLCL<6_~xq*G@>:!;'qjLq#tvqm癇۞6=YOe@ e4R#'Z\ڣ4/_bOo|c @d8UR/SX<6=&Cov?_T?\_sEn |4rr o3-8o]]; ۀl n犡}PQx=PAWjt;Kxj04 g*ih0d\H,6jY!b q? 3ljSFdǁA;nh<m@"Zq" RR*!2m%唨pT۔B]ضh: l?;A2A0N TJUy{*EFA.x W<Ђ^pQ*TjSP pPVjN fU y0|J{"pq ^|\`5Ktg 5o>NȷQ6r0[DZ_z+Ӕ5Wmw Q^tﷷm7N~0]}G⋿OVovf[_?'۠E]"Qݍ}J^DzDȬ BP M\M$K^|ws@% ?WSbP%̅h{sp-% '"-'"M钫J|='%TƓcPy-U@iAVެY\Z&Bf=m wP?q-AjnD({O77 ڴٻFc^BC~v7Fx/6=}H|Hv !zs鈁ؒUU=lJ^X.W l!t^1kU Bd VycnRHx~0<&L۠ݤIJ1PM0f ʠ+cA tIwVrg]֒K;Gq? k!*V!z|ƒM~2z@#Nhۈ6 %yʝ/wy㵜XP!\ffq;qCȃ2cۇ3OW2s2(f]7.jROw Y`GTN:EW:=i2DŴ A56CQc׭ l e3U9Q֨z8*Тx\?O/ߥI|{r=;Gq?{Hl-Wz6—NӲlFqn=!HLb`&՝EuVY}Y]rq|UT ϩ `1IKCλ{% xOkL[o ZviRzƕv{bE9RwqŒ3ӝ3.*d8]kн 0LGfݝ@3bȝʭEɣ  i[}  Aa𿼗Akk}uPJŔ|A "My&$HWP0uJH W{X@z{P=:(Qϥ)1; x<o圖=wYdIP&xu+bp4Ul(t&tiXudo/磏RZVTL7 ?߬oVoN+06FSt*X샅..Mtףk3+0fx}}%ʘ;c~(y ߍDsNYm߮՗6h_L 6'e7A Bu~`VzwbGW'meΑ, Ϡ0Rζ0?gMMw}͠Z AN/qSgL7 j#FCc HR@aء#ETXʊ 'Dʁv5:J5j7ZKUy==ȉD--]=-$-xti92Ĉ?J]!h uxa{n$3C@`7% X@92dXsJ2AkmkqI9#B;a 0NrcQEp@daΖ?*#7]y!gQ((UXUێעZq΅цe3cnYZ8H[wz*V rCePkyt~Qn S(,W͈Q,vL_,ޜ,m"n(8?z3|ט_Zon4>l;-m5o֙_ѐp)AIvӤ=Z\N;XTά[6֭ y*SG֍An2HQwԱn}f[6֭ y*SYp|<՜#JX! b3@14wq.%YCP މd@ 0'(8_XD9P/Y8uiĨWF8euFd͵[-P `[smI -0.9jʝeIfԛhoGeqx(^9ڬ.ߜ}68ۼ,$Mmޝ<6|ZO&hR&i͋b>`>h1 M0jZ?|73K+fDf(><}cro/׸*.Ν{jRT)7\f2IHV;YlݹJ S\WWwWمӏԛrr1*/.,D6 Kh/ @_|G@e1 0̚$5Rt.,',M^!(MQٜۧ6DɁD"էG.$R& Qsݛ& QQÕvE3K|삣n D{^!;faV=bWHhŰF}ō_ڻorTOt\) ލW+IՇjvp*cWO@ Y)M@"s4kT@)Pܥ+o7j$GA "H6x.qY'Mt7,?>6/bs~ONÆ cX ՈQ-mzqv_Lx"d1GZ3/5fV*8 n M":GCq;+RSBzh)+y}pA9 CA:Djn#z:'gSEhi^Xomv]&#ZuMbh WOb}  ~?Nh8ڧ4SyòKyp9kN(}r%#JAA&2xD֋"{+4w[*6 !NS8ť).OեǜIlt<@&KL)+%-Ƒ Ԏq>z]p|m:.v7~fJ5w~+7v1y]z6[m%6oEvk'圡mbor$Sw'1C =,0.BP-nP'Vv`ggB2i3D b =d,,e!c@d%湕Q_RQȓ5JPt42Njӝ}9YSkHޛ4z4]NBX +s}Sކ;)D>iC%Zm) FZ*D >s3ؚA Z]6A& cB^[iC.Fvf8?dgc"WNU|诣{>7w/~C~ $/5Hwt#Ęk-״t:lOG78>h MfzOE*? ~o? Um{;n> a d<74(\=6&;⽇:lwiL ỳD7!(.hC2w>6`V_2EEUѪ$ɖC@©+(i>]TTߥ r߭m]yw(r)~ǻnC=g\R/oHO * FIOrgv~X;a&F4B9P \0l?CUnsRZӦm>ԩg*'3HH; qo@-9awޔ@ANm9&+%y㥩tZlW ֡ëy9m]!X) z"\@^b൳pD0Z+50cЌAdbd#z/oYk@Wk/4=- XVqA;n!#VqD_k횖Aw OhOǧ Jk5iqD[5i,@/$w655lh ֚`Q]_|Xmy:Dk0 Z @FBuM_Ӧ5wwOWwov6P ~ls%Jutc߼CR )CRzAmf7Դ=Zꯇ E3$h(ӛ^B?PZ0:H,z7ҙ"Bʴdޕ67r$׿'+V }0_-: 1ٺ+m̷Cp5Hَ~Ӛ?gGCO%d){+ƙ`À8aY+j=vI@ )P4HA3P, ,Y V(i%@&:eX[gg%K_TT;+3ǂQ}KњIRrWƢ :H:H:H:hks,RDT8eDR(>0lX6 =!9 rԖr&]X,}QX,}nycGhbCcM'fǵWMr_:k2 $)Q +Zd LЌz3w[!)U3;$iJb"6 +!9>BhR)QPUpW1iLf+q5{m8v.e%kTy;pc jmt~^gbtYdǨ29&sΡ-20歆iMp(͉WF'h&#B^+iKT:0 h&Gm$D( >z9"\)6lT~ ךӿYpS6 |6TD # [`,VP`H0@0l'HU3-h /x+X#9h)F`PIѠkd`s{^22(Y Q0L0eՂA䨣TA2.@}RSfdgdl-, b0A9Bl-rH*jh3B9F'J05~P;wtc fhql:2ϱYh^Ə'qܫS*DM{w"UFџ]o]OO~kjpUCyLg{n(x?"R !|wOT>ǚuPC3a|LEлO B^ M5 lrc%ҩ(Ee'q 9Е mi/Sӫw{ }D1XJT{2ւ*H .]pދ-ũ`gphEDaN9`+sHC̯2ʺ5Y.kAdIb]ale@H d l8ݑ1eECY:T e1|%$%*E"P\삇TSAy2d :F"]DsY@Y lY,Qg 犇(u+J ŋ~o&[BEz6Yn I¾[56q_Kpބ|F}ЎLu&'ӘN*&NRӹ4 8ieI,]L͹dv|8LNLFS;h?WBRvC)# kvX:I b{V<$^%Km5uϤ^U͓'S3zʙŚib$@F0#F^X,c7FRi>rC۠徖kփ'H) `2ao0xǕCӇͺq200DKf;P` d5$F0qM@0ui<My(A2Q+S^~0@gCWe*$%bGf/ `̱sb&/d4J^ VHC "M(Th,ҖG7QiWQyU^ >hcC 3N "HAw>qHY؈< ^>Yײ1p䓫; OQ])?ry3Z:\D9]]\IxͿ,_3 a'J NԆK'i[FuC-qҖ(eX>Mzkk/-a=$o9֢JHM+|r(tPH7 ۫KbXh, Bzϕ1X.7Q24\D-B2j/];VǙ>An)R8ikDqҶzq[̴֯; E:'"t-q޹K\X\, s6Ym[<MIt-96]M`)If¸cW~ k&Y~O|=?T^p`2V1{;Nrߪ\4|hTWݏFռ4U/qnijOu >?Ngwf1U{<>Jp44Y*mw4agXEޙY{Lw0J9Al ߂|x}~Įp? K n.Ƽ6 #w7ޟ՗/at=o{GbFpec'&ZQv>7}K 3tD_}poK}DcrPD;16t榼$5]˕i}b#lN1mzU)&lqH]f>pZSW.1R $9HƷ~ޢ%~&T84LndŜQoe# ()F+gҘqm#NuLӅ#w`L˴޻4eiY37wi^ӠF[Է#"Jw8Qk.=״%A>ѩ Sh[lIM1婎앒$'v>&?WA!ʼn28^ᠺ#-%a9^Q-]>tO$M;LjdCX F:Q'-`WJ62+ީ52ܺ‹01AK>+Htc(Fc3mTJ82kpG_MLll%L4a^(1~zf`Hv lBq?W.ll2vuNatP3oqX!ީCitz_oy_T?/)cdnU~(,/::::&J0V/Zj@wbj1TZc)`>qt,E7 &N>eqa/EHZOﻋfr9ū/xY6ݫg FުL2d8Ve*S$YTbT)|{o)i(flǣGi=n/ݻp̯YD2Gk'.~U]#'J>nQY9" s ȠvZxS/ w@ BT :ylzZno(yM QKdF%̓a5c鰦@+GJ:ii͖xf;4+Br[TJD""6ikpFu6:X+#8\IE4Kh|~%^BckJl9p QL !hCBHe;7dl}<ӸZ4o @VqzMUxI>bA>7omAV-ro \׻`reY(=[\l0R%ն=!!4K`H"⥌m7&ָ/nҙO7+`E4P̱D< 1 >44B!Li y?>UrgĜr4|yCЎbPt0O/ b#)iF^7Fк wFnuhC^Sb ;xtFj"(H?#שr;p؀GcD1qԉ`1P0+尌>hw8s_oGV,2E: pƴ@"+1{|í].љt~M$ 02ZfRJo,Q`^ ַKK&쩻~Vː`*$H f!й0ԀmR> co#PǙ>R7P9g~d{q8 [o[77(:ۜ/QiO%A'+֦H;\Wk9-xS\%m#$?&bGhr,N 9cAPm #R7 EKYk~J9~Ŕ\+2蹏ɠFs"OgomZÖ3kN#I)k Z\]QL +/Z}ERD9~YQoSyÉAP3ZQ1EQu/Zᇏ6W`7>v? q/*V2^ʂu} :KRK˄H7ᷜ(-LLy\ 4Z"Iiej)D ] )BjcWtQQfV8ɧ0'(eC3 ao{Oûz+C0ٻ6#K(a@8ll'`s^@X$]_U)j'pbMw_UuuWwcˑ~ 7K廻wX-(B5n;LdXX}w! e4 p \12ǖcm4.|HՃ)Fじ:Atѣ\^?[my/kKG(F][ogs['ZTUy4=8k?YtM׶w^{q;ЊIŻiv[IEwNVjqU8s>;i=8g!Lur@(ED(ҳvfG:>4G[sQs}hMgXdŎ+*V/Ů=ll@t-[nAe}Z援] 5r//}`fˬ4d(뤃'=<Ԏ=ފ1MAk/0oʝȫN?8ۻKTeT"Yv[ncyO9Z.Wvp{N3f?4ނERP@߽yS25x2Y7D+;nMF\*g: OVQT_F \9嫯ci7 І?|xf3>Ql~vaub*W\8A4jV)Nr~oOr[4`U/<۬pJ:ghha ql!W:{9 LB,jСw)4$|Mgoe';!.~NJ `p'f|u2PۚBJHsȢ4o)&(jŧ:[`ue]ZEO˺T>͒|FSYov7oyn(g[y&wwV^!ϳwLbfRY!X7Yv Ѳ3)Bj&eg劓6#l5W_KbgG'͑ 4,>l[=i޾S=IB!Ӻ"}!X&AQlGlOe}k>ځl-{bՀo"=bUo^:Q=tiTd܇(s샾(g#Q\̊CNSQ1><$c\ytAVOA,hTVN X窡vr )rQ:ͧl6eqh<+߷f2MUA]ѻ<|x݅^JwK*r b:ʌ>&c s'Y9ӕ>g2ze)B3K~͕S<aP9BF8Ɉu܂X!p1 F`1ԗ#*q]h0`e#\eV %" D3yΛl=Êh9e`7GPc `A`5T(̒$a A[.=50HB2"&)\)dnB ^h˳NnzyV>wP1'qвgf{x=4|Nn,ݷĸ/! ci/l避QBL52J/~K޾yi]XzD #@Oq4h Wٟ\!Swnpx7`Z_/qAJ%VlsWj1axs,Y>r:[s-ht48'ڮA= TųË&킷R-@X;Xݨ-ߧzv"_W'Ytl>L!5o'˧~3Lno_0WO&<,SH?e< 2}D1ı&A$d,ha e}+1"Aٗ7}Ϳ .j+0#*=w.RzGP;{f_Rgƅ  8Onj 5,71.m{ ?/0ϟV"UzN4U9#Uh@nqt T"I]ώ+LRp.|$sSʼn_Jt[vȜ܇|? ɖ-Vaua zn _og߆ Ɖ#J%=,#萒(1k/}8X"o= Mr%͑~_d-6z7{KUH}*+{bW\Rp,6Ƣ@e؆T<]&ςT}ɶͱ'yw{;$LTʻ5+ +Α4w, +g#M7evp?5sM@G?Tuٻ[ܺ$]]ap902evb(p`3fXrE wuLS,éUAC/MI_1+^QO`Š *fut6F>Xy.:t\0ꮔm=rQ"8~"O#fH^8jI)r>hE^jTq!s+_'A`,{ dP-da*IjTXd+9x@R{~ka6+4bLc=^`dڻ l/χy*͋Wi^JU~^\MF7ȱ R@ |H#BFFXmZi43^E84l}fs&av7I73.)̲6!h涎|פ#9U!QIBZt;ÜѐbtTPkL'+"JO@aL467r#җs @MU\TKQZ{CRR8Ù!Em"x74FCpJvWźPVɵS26&HKxH?F`KR*:_9Bh_4 I  u9gt1="2.g.gP8go*fGV@4uA&=ڳSJ }ML zT 6ei(Ƙo *"͆CͩFS/x;˅#&=˙\Y#+LHD,hD a-bxMaC0x}^-.Vf뭊_&gpZn&3:(Ttgv~ v\O5?;}[!Ûy.&|w܂!K_bZ_ Z폿s2ٜ~h厅^`e&k&wl_xQHylv2*f=3;sd/OWYt3d I_T?!8ӱKACߡQsPp-fӶkA+P_>2/T_!1|#-N3wyr1${ި58_>Džp3-=-9Ϯ?6O&Fw{WwܗCҿ̦7ki8_Lh#Cl-'pʜ׉+(V?]@F'רTKc.F>[XO/wfv)oL*=lLRlb'~tsBb\ZO${rO?5[0%{CcVlv7"Ӄ@_)77١kHk}h}ڠ8pl٠Ki&(lq㚞KP\yFMgg!JY3gJge)9PeRFc#oy GC̤L .vգcoT *HG{1פsN#:fԔSr3~u1`m\72L׭J{YF ϥ1cL sE5XЌv*&Ħ͝[?Y|!IK=; vq9C˹p|sA4LpM2=3CPf))U&q,v =]͹LWw*sW Jq< JĞU0Wyf fW,Tz%$RO^^O:JQ8=}WOr 漪~X"Ty6dUl_Ul_#6drCQEVyf9&J/ZnoB`}a.z_pf5+:~|pجxJ?yt6}ul?_~Mtn:#wlzʣh/Z!,)xp_:Ez.On<¯f.\^n[x(>ɉ pRܨ^+E]R"ZFFu8 hҨBr$p ДV"M]x6pi=e@ 3S=!2."5jZ|w.̘n:z p57&=?-Ƌ &I 5OxB(Dh<'"|K03G V(xgi`2 G2$0_]$J5o:0 ۗU'Is?X^\\JB$(G՘6@Ő )jZ1Lnc)NG={tkQ: J.ܨ"Q>jp"f5Y* 4X Z?v'G+fR'Fd+;';ݯeMU.'zWVh9ݮ:t 5iVT}PRll1y;؈26hbFv߭:\MltG*C'Xu_}A^=콜wʗ[!M `op.| aJ'bE1.܍}z}/?hJ_n0盐朤wm(jYq^ެ~U (Xs|WSqc6Mzz:Jz4O;k\>AD6Sf@2V p'@r=nxr-[vV+퓋2vI+dGַjT 8BH ~ԧ9ysb>&r1Ơ^o* 4xqKrq|a03.fiUۉH΄|' ?ȅSK?H(&Z4]_;R7K2($xm:-f8\1˓{ݟd*wryW)wryW\r7drV `jJ)95٠tnH0S+ :- ~MyWqқG.tm@;~P4/?Ҥ9`O~XF=n9)|v~yu]s*kW\*keAV1z jNC^Zj$0JxDJ TV9VYNHN~ +wPi'TdKya,^D5 ޏ~|_;R Dg"|sJ!ҩVew!^boaaJ{p47*4Ka8fO _h!,_a*bUz<}HS4 A8U,J /7>"M (u`.さ{|=*G \g}N>(ZB 3ʩ>D+sk#^#|[NxdqcFFJRs;ޱ\ ]ڐH%k40`$k0^jXD%fZmCK0ã6] q5E:5wW33(U2F˩at.D"I\1 ~V*I/Yd^7,%;oQc]<ёNDOXU,BE&DU奚mf `eF9Nlxy?YJ:NJV:>2Ox6bJ$ǿg=D̈ Yd\ߢ^vnP.o.6F9oRT]~6ΌhcYNÍ|0LhRdp0bH".e'c.xiL5DUѸ{/5{lR::p36Cѐܳ҉B/ r]r=5A mNRrؠt8Qư6J@$U 6(dcrBTªB#FEg жE^$'1|F(h.n|?՗gAbx$߷on q[IG7l٭ڰoc_wҕy)0(!X1bh L$ɓƎ5o/.煉/2ҕ\gY,ri D#.`# F5$h`&87L9mp<IY)BF_r>bDSWOfWQaY)SfqǛ=Φi4b-(kWB<'Z-8f Jb(,9&~_fs\p?cE_2hQF?ܬұ˛3rhu~WdDԾ%*Kn8NŅ2k5&',y}yͬ+Pw%e_O~dgN/ޘl=8z*pZ)SĽ,{Sh͉򛺿c3d(#HDpprs>!Z !\S'5ҺVܔR5ݿ[VsFN[+O[?~x,"}tb`KMi~z-: CmPi/]HI$nHJ&pb DUAyDǟ#][ A5L=k2ƃ`  ' ŮDA$(3%4):UsF nԨQS56Eܯ) MD[IpJhH:Fa1,WwF8"Ouu:€`r/7O D{%Fy))i#}H_0) SQ # :^˼Ee#8L1:q Wt5 1UN(.9)al \ 2`E+11fw Q zނXRt̟ܞ_>|hbQǝWƍj|1 rޠ}d|=X" qo${s2c9qf?wyo5gfdjzrZws+ $f_Ծcي,n-Q%k|<%4OϢwR1W:=?Y/~?מ0C5,E9Mu!\Ekt4 ȏ5H\n([-JTm*ͺUOԺu!\Et k֍6hjyP:hc?-$YꉖZ.4䕫NANHƅ۸w0T kuH!ͽ(d]NBjb=G9Dc 6 *#NEK\L@00N N* ܠE/ Lˬz&ފK˭WBg4tn7si'Bh*/ 3DHOfnn_"ƴ!/R_EMsQF9*@.ڍ3=Lr oy 0&G~ԍe"3X]r$yž6 cg[<8p-7'ҝ.^}ӱEhw>y`B1j߃.'^KDZe@=(' W˨ ȢDͦN1% pʰH#D2HRtIˈvecfSKfx-MAz^I~"} "xD_#I*n&2ى'_6Hf)ٴtQhϷ[S7Vio݅[!cBCQchN)2Hnh1##aSmjt)}޻!5p`竿)Жw|Z"1m]r4 i)Ï fp!gM^%Qw`Wkd <@N[HDmݺ^\I#wC sXnmc=hVyqT{V'1D>? ^gyl%]I立7褦4ޞRFGp^tpiS;x̧D?}T[ܬfkj8NC7_T1Q۸d-՛lXYh5%C4~ (7TZ %kG^codN͊, >q1ڳ!Z r9zuz. =9\^YA#ym+b n]Q^mC;FӬOA2Ł azenwh۱Qs؎0͈O7tWզ|OBkPlg" D7rqW_;JA,[͙}_Ee!f$YýI8@h|1ޥDH _%`R` >j8ؒ C0!{^1ciJ^12Z.g@FP䡠۱p8=r˙( uݨsE7CLѳhHY'k.YrSZŽh6բߝj$zW jԾݩpuVb0w2F&x T'rY^3CbI %)(h"6 `5\t[!U-^Ј_WѦꀘ<(i^6W#BL.4䕫hNqqzͺeȣ@\AE;Z!.RUz+Х+h+W*2#5ƙ< V˃թ}Gv]~KÓhuBC^T$#BQBg|rW:7ձ`:c<|sx.%#lsi1yDK{~D$L 袬E5,ꨪwsWuX<,AA%8K/nXn$_ FntRLFG%pm$*0_e \RS \ eFME%p=}VLP{ ܮ>y ^C)l,RUﹷ֔bt2 Yض&C*_*946qmX߮x;#vZTޯR-xဇf<(9}GCN%=tYUJ'Ynd[-JTmiо[}Bօr);҈sͺI1XZN;XC"$k!ۺ7.n]h+W|!Z}t*&GLriXǕMeY*JFVTuPGɞZSϥ !] &*Q03\6*:@ީƭUl7ߞ.&^n"٩Sp])A$R\$C2Cd)İ.Eb W"fCm@쭄!M. FC(8f8fBǓ/Mp$lQ t7Cc$!Ml_1 mtH.]EȕpOwH}{Ag:z;C\'.6 7~r:/fّى2/v 8t1C+]5Cޙ/2χeD͚,$X{_obU2}Y y!`8L!.JCgK[/vdBɫVag+,RȨM̂nO|њ=C5!rᲹ]?p|)&1;_V4:A;ϳvgdVIʺ$f%n4ڥR&#>2d|a]R:l3NΓ3B#0bwCO?,癢<NhA4~L(n6Ddӛlpx"yFVG&͈ uΉQkDvsIF7Q jĈ'Wz.NK?-Iu"(K\TjpT;"& **3l:Jf=\0JчX]ܶt2Xr||K]^XgR^Cu>UdXYOZ$Z'fTz Z=xGV1$.m uYD=rQۦJ| MNh $ t:hmoGՏW6U#/kİz :d*xi ekdj:<H 1F]t7'>1|VE;Jŭߞ| :!75!8UrB0C nz6D܍'_/1qɞW~˥%DGgM)E`>^?=,Xߣ!<]ebR^_ܺo4)ɨ=Ok.^}[G.ىֈw54/QN=#JrV\L]W jL ɫ=].)n2呂f *Ģ9Fq҄JWOKqS [xuG/gE(V}Ԧo^oMޡ|T+O$Co&wK,*%DT#0_Vz ׽Hmi}Ay,/N=BAr: A%`&,>Lⷧs)S4RW^P)¨bRdv^V[oU֡N˻^j4_A_~R<+w;~`A| fןkС f9 *;xjzZ 2ß.8ڭƩ9SO̗o=$LQNF30|6@<2JHjYqҼRDecU1Py}_xESc0TGl:a^d&tD]5x9[),xryC)To>1d.ĭ{%3I*hb*)͝Cm5INf*P7E> A. },'oջL0IaV:t!HήtѢ1OuO4-ểbH!4@8YAqcjHqef,;h4+mNk6Cw,s KyuQZF`BĩfY~%}vCI f:/+_hjC/ 2S{$g"+Jz3- rFEri0c&::/&rWURD8\żox-LȊע˥a kQO|/WW UBV*20"V J$XVcZkEzh!5WwIm[7ʑlj mȉ#&01(ZmvvS2NJ4tµ ~%d^0OaQQ/% uRqs-uqܢMj\Y%vOV%Gû+"|>7a>ܸwwywTb^uɈ6ri惾x$A{/MQCnn_ΒJ{ L=ҼY_ OFw}>}͐|2\٦Yڡ?wrI/I$R/7PMܭc۝,vgخj̈@A3b̡Y`tRs>[ΓY_s]jzQ&EDN,'z4R]EíXʅЈbB2rӬ5E󬸻:6s/dFsWNקtӼ򎧫w;Xx|0Fob IZ[ҨP)Cg3:zM2i:hc )j{X+)waJ/5]~+a/`LsAy! s!GUz+@RҜ|zGy4WpfrFvnw=3?-s| dC+ZYY:2+Ǭ !xh}]JA!iwKlL4EIQ>fC.r3.CݭgfLEyMr9W0ik>jK)f.\C1ha\X{Dڦ u0ܕZl*Esf-9 p;H.˲;/kL5<ւB,[C rQkMkz{IB >֜"M>ۄQ`4&Y)Bbmu!6Az6樘h-z*uٓv6y@uPdk89ViaZƥР LXܰ0erԇЬg7%d9  '"(0dՓfn&8iN0<4F%gq8,2Nf(}Fo_6J˩4.لNc_B|NS<߄{B$#g&誈B3>d(IgNZp%\㯚Ixo&&'`ؐBLbWշoxdXՠCڊj&4vιP+UyvRncZQ_;7͚.t?"86CBM[#Ճ:ΧjfsGFRUXwGM+8Lb  ,4h*6&^I&LAw'= È/cpt~.ViT1cXuZKt=b[_L{ݖ4_~궗8ZC!Gk}5DK> 4HrUpRĜigdžB Q7M,ojgME`X"xEq4C"KFT#ph).zjº.E}s\+tϹb/Nɷ6*UJJ[!M%]͙dFϪg߻`UMk/#-BVOl~q1'7_!oIA 9 `#F S6}X{@i긄ǤB q =}tK`$Ps[K}JN?JuV֪})!̂B!B|x0sKɒߨfbq@!ՖK-4Z P6( r/pr|A d7w2tV;q})r*b5c(v:X)2].O В4 :}08K^sȂWnKPbUr-[vHF</ȓ11YreWA ` =w*w=nHpؓw۟&.df, Jϸ_bYn˶M-;t[O,VVp1:%fw}c\ш.&Ǿԭ@{ or,L&RRMh6Z59^FI)5ˠ&EᛯC&QbֽS1p.G KWo[RYs"a!,i,@ jiC^B+yu蝚>k*;k?,>-=8OUy3UzrNVݫb^bOdQ)G`kvqo}ضxUQJWL:oP1)8ڒg_pg;Eh$wP1?=._NhuUtZ,zvQ՛ SCe;!Hxbt|v'dۄl܈G$ lP͢:,vt9PMګztT3<[kl{qYtq0c6]"t1DK9e0G !ںVVKRW@.ד4x*6;Y6g1K%s[J-hzF p k;T !Zb2FVP@y󛸼yT. Kxx}EgUaC,LgÕW28;Vt3怜y 3/-(I)f)ƥJdv0DWw7 e8eyVN%E KMm-,pɎ1vr* !w ;R wMSz0ǝmf' n?''y ?Mܹ^'eӾӴ%Eb Eׅx ߛ4 aJ'AO rx8⊑_83|ń1NӶ(O$o>Zi0˂L]բٖ@OvEqH\d\o/RP?I M>{xy{k'$dZMgkrU!qC" 5:dgl4b#6odьup~aD#~ˎlVs B> 4c`ùX:8o?_KG?\F J￈Ŀf=eŋk sR, L_^3E?dfCu>̼Fzo^3pKK@SqJeJ((k=-/k .o+/zͤ3^x$Z׭JR?'&LJ5$xNY: +%#heWfEBaTw}Ü{ȹsn\ nzљ>>s"Ӽ7rF[ֶ.n=T0{ƩFsu#Y`d5!)\юYzIޔo@9Bn@ RT-T *٤RZ柃x_A~yސA-.l2i |' `Fy70ԹM 2R 0B@(k G+.Uir.D!J9:b7 M(LƭSf35h(1/?A5]K0q{7b?7\#oFp*řA\<\22ܤdJ%9i,54Xcb P~R,Z~|)~Gyx=QJBMSgRQ&Y*ȀBΉb!P(]j]A8v()|o)h*bIzPOEU_,t[FC1!j( BABu3&(H;lΊ49^?ynl8~tiy1$W]!uʱt "E.!cH5(RtmJvd&֦:7`ͨڦuQԦx~%&9ߒ⛋{l :qaj˳Ma׿LF+<(Beϝ/T;gSҁI!}P&inD:v~r"Ҋ2f8Z2TPr{6t_VR|e^*Ěh.;pBSDXtj 7@7Cz[ h2yTR+҃< ОoYφs9˔*w2lۤI#oFUT͔Z)iftJv@'3cI Δ(*S*vWg*\%!S[:w~TyS1bT5U 5ϩ/rʹ.c`2DExu@Qr;ra Z^? CAbu!Hv^:!WBTմ}g8/\ \TYoV/_>Nh_:|;H29BW[!W^Ƚ}wDN0C$7}"[JUcG&?΀cv|9$SߙBd XBEM&g:IɔhNv3QBZٙ3rg"Rri4e:{TBJ*5ƭ {p+cSv Pn)mm 6Aʑye Otri;õ'.<'6@NrRiGP2 ;8\mNNIWWk."̈́2 jF Ҕ\ulbF&x_6)7[#o1խcɸHM6#% `ʦF#ÝN99% 0M! :(j8v6k[ǒ 2TQ8NLAJe.tY8Ӕf(QKPku2}NЊ8pf|SWc=LE)vi pq["+ l>#ӯta7F97:urV+ NTJ8#cW"(b_0/!݈'>LՁM)BgE朡kJ URu sDlqDi.EXe1Q_zHr]>=pfn!>[^pe m$Wwۇ/@zRW|Z*1')BU> ?M/ K>|/U\w|I0j_!Cz 0e|u.9H|*+P!._URJ#^g2`Pow.y%t5ӋqC>[zښ,Q*TnjFqUz u^T~,eE 'q(bbVaZEMP)۹V=Qb>5эե(y4Yq$,V0d$1nI_m;QM_ƃI8GqKJB(˙. Y*3\DrPDp2"Jf5!À4A\ ] ]#.RhEښwWԐhH:yR:f:= t y}lk0HFwmm"oMu G l)A(F I+J%R'ilH݌Tl\@u8W\KѲi5`7x5+7H^iDF iߤNI\Ж %{KA׭ 0֋ 愡$\F6v)!ȳy `pN>w(s?قal]"#;~٤cT.kxvrOQD|&w`yV\bܺnَfâ_hᡎ{8w[TI THOw9ʹ\i: <rI8qbe!MO Y5[ Մ4@cbaV2L,2bO9ve4@|Y:!44BSlMC4+} fyU̯oT0FfEP>L@@7M !cI '=>x)Wu( # e tMcz1D{5A[,5\@tWJ"m_Wλ-e^S^DK䓏ʚw. tu1YIAjbġ?XZ9,FcśԒ.:);Ƃr?76BGg66<rM>R6r:.{̍9n0h !=c|^a"% 9s_@Л;M(g\Mfkw(|?9,r6R) 8ZlJ"n[b}.76MW_r2N/C_z0s.co \`[_@<ȥ/ʂ/Q.C5 S8k M;PTmi`]5V7W诐$eB-AQŹJEnՠ@h+|1<&H\qRg9sQ͌PeL@iD.1w(lV 9H2l4ZVGvZU ZɶGBm˰/>\>7JU.iZ(LL@;y+c̈K#)p-%~9!W7pwѕЊ4'ů)]xImjAq;ҡt 1(ML4Vڱtv<)|ZY{na5(^j\N=DtF(t0r£%FtϾ 3Hvz`hJIlX.Ǎuc>uft H[JIQjo}T,jmjB5 YʌκZTJmkiJΓk4{ Ds/'Ej(gֲF9h8 L:zmң;@L~5٦}6ՈQyKFlݜ^\~XߙqlSTM%wv1?X-]p-n[B:'62)$S.SӐVPoYZ6 G Rܩ򉍬j`:r9G7+7NcM~a3\z8!(fM*2k&/RL3#FF-SBodjW#> COy&{AZ@Z&.TMb%[jَ!ʑVC axv\}C3D9D0nErͩ:[卻P`pEQNN!5R(.oJM?Z{ 4HH'rla흙']Jݠb2;+Wx~&Wf8{*F)~ZåЎx5ǸG+9Mx$ǖ/qg5~*?zQI@&:05^Lf=8ٚlMj~yՋ|E ȜM3*\,EY)EzVMx{1}XZRw&WߝW/L Q5;w9:Z~u5f M|/8\ c:ߵJ=/R~R H'itb?o h0}'I[T+ s'4t@i-4sfDڂxL%e#rG:ɢXyѲg'!zj5"Ri*hHk?Ts}˘܅=rK|Pn-r !c5CJ @uh;aG( k5ֈ|Q&7wIŞ'j,`w2 9:ȒV9sQsN˫ -}\^@@~YbMW7n>yzu s>_Kv7_1w04?#Ww*̛d~|fU)~fu]?zr2~knlҺ,{zc6ܶTcBq=ܦHF{eĒ]#^7 GSFk9=*r=h5OjkX*d:iO7@r4%LTq*F(M3i"^L%9!r#CNMi&0u1:OlNk.N3ɚ<;tuj*rTwK3GlSK,:h-'l&7&\2&+~6gcŵ}ss~yaϗ>+YQx|ueriڤ*sb[OEȄlmF(Z=pS\@GA oαj Q1pKX$9>Dvy^5h~ЈoZ|\wus/55F54 Kx"{]3v~-w4ߦ_Cs{#]JR6^4,莀A'tqLNTZ&ZSr R{] ^xW#hdG^v(Q;ItW>ޙ T旨thO +FƟ7} YMs` 7WDM(G2^em/N?@ CJ?H UCt؎8b OgbtxZJN}) `r:2<I=C5\~|rL2Ct4 ;5~7ZV^HdN{1+[tyEf"!=q' h +]J(ۍp >xhKonгb_zRjQ#6SXhz7ŋBŋ) 8VC쫇jC:V;c9841tsӓsR—*'-Ho:+lpȕ i*>Tt"!$d\۸(1Wned xdK,/!2k/!m2~ Yi66Գ$dqr(֡&#D1"KV^C uONʬE|XasF\s2p 5&$%D|p&eCI֖2N%6n])_U#njmm=uV5 (&t>T@N> >K@ QbTB v\zl_6,zxWwh &(=|=.Q<= ]4 TJy_O8MK*s@ 1ݢn-d; *:o-k8[ q놳l5)l)cK?7jnK:xCtic=$Nr%lDدr)1ެtl1>rF{1MG b5> ǽ he@ [pw[lCגP8RBPR!*Td؂"~$+DǒhٚyM1Ή"ip_dM #}!5&P i1M&%>dlK>'r#S=Tƫi*+cmEoWɹ4OGR=RUiRQI@=9C+OOJZiJ꾙OӔ靶14X;IE òMD>M$dK$Y:.z (rfeBHf sq#Ww CU:Jq6uubaf02ϔHǕ}3Crć qK_V , 9PBuw%آ[+XpRn*70O,-좘cpNkPi&P M"ti-av١/@#Ȏ8%b@M@|E'I= XD:{O4+kѫ>y|϶Il## !òFX  mq DOj"]GD oKGJ>^ox9jfO[ =H3abLtiYih8Ұ8DXXEic)z \w[۷ԗ<2!RD$QD:BƘfE%B2߾- jVTpu 2DfR[UUV,X I"kFD6dqC2'Ԛ[ꙅBMJN{E|5-+C[I/еyZL^3QW="M3̰\ kG،gz{ &tc#lbU>" Et%C=DžV\!uZ-yAo7T'"~[kf=^oFs40쮥YBYϊZ^9 }>j*W>J]wqRS$G dKfo%iR7Mb {5zȊ*kx0iL 0Pyc9v8N^HO^H2@& IL1k4IӐɼ^#Zkח bKd``LJL`K U? :u\ѹ|xUH2'e(mM>D9% =Gu\ZBns+P$dD~Rm vSv}!x)ӏUFljBn~ D1Q{i [^+hl0 ?&E4RFrN4)_bIEա烈K ͖>^0 XTvz{-s?)!;' ZړzL +9>;sݕy$ w5k JJ\ c&ؕ}s.7gf>$FAm8A`!&*Qk/Ql{|zXPHcٙZ]ÇCu‹Q[Zi{rjr0isعw q/MmN+P}e$o8.IOcvySi>PjOxKx,xgN1ضJQq Gunm2 q; G.^k/JnRM?OHB(T&\p㽹 E eDL.q%rx\vCd]N#=?QA@u1sB>͚~|4mv< wh<ӏ[a&vYt%>O\fEyϾq?upʠ:v^u^/׃{ĩϯfp7<'[u=PKrw}EbS{~.\2 L04`2tMG{$t.@X4GwuޘwIl';嚞}z2޿!iLxǯ>Y8.LB 1lL`pQbc1$b@Vʼn7J #m\HLM8t?$sOf{n}8yRrm^mdluf¡<.v}w|)DB6ji#aXEJQbwU\ L@90%I@YIʘ0!CPpR(*x,4|R`bLZ[Y*mOBuke'e ]+W 9EbTWk0ROiuhAP[0 D,5'*FQtx".,{$_/aIZoWj JV1JLRg#D&,l c,Qچc jFt2OQb AIVbK!Jn"&UpBr*`H'8dVa&2ra["[@|f^$=Dq)Sqo{\;cvZd[).~3oDi\݅U/F:Kp|wogt?"A1;:69GnwR3D.(ݹã=@vf:! Fj{8SZB(rܞ WWw=cP Fsx76* )/}%&Nn<i .+!yjsǒ`VS.!yp\m<)CQs@oL9%PcU=*1;p#xZX,| X׶ יφSY=w8Av6;lskђNLN`:|g85 zro<-z"\|n8|"2හUVe`s#UY$jE쯴kEdOdR+G%V[Q_[y\\\%3Ru**h}9*S¢4 J^ߣUkQpt'؅%A b#y;R#AZaXxO#n*uTZJD, }0-00<`UȄ,qS|dcT2il&  "|B q6s҉5_pQ8- DXz\H;NuLȵg\h&R\~)@&l{ iq.) CQI)ad?UR9lЧ%Մ-IbC{KEx= 0+:rtܕCݶn@5Ԯ;dzDP)J?Cdjhh2~V0<{ux&(e.72y*a[Ӕ i֜l~l'-mYr9nu?ޮNO}]uKpTKHSlYiuΟcX6Z,%}PLnU3O7lmpםdzZFV]IѠEB#X1p4Mg@A =/]GJr ! 8lqZxPb]Ȗ%udYx«l{ ̓e`P% `3uΖh*L!-2!Bʅ_(:# !mUSZW#];|O\"P $wK)^[XJA/!Rqp+laVhelYU]s tIlUHY;T4D3`Z nt=Kզt׵Z-ҫjN5YPIRkg˓f_>K\o<[űD`yZc99|hJRs3ɖjE[N3S!ut2sA7ŇM끿 B]$e/)[K$OuW'w) =eyL.A*$5:9sJ+ x6+n3ȾȾȾ~Q m"JbbBsFyE!l b<1 ĭDq](ſٻFr~ e~i`r7 \=d_0g9O%ۭd<ؙZMVAVAf_nS:P6w#3b0fJkWWZsSN&H" TX댕́4 BRPtF2H k=kd=XH Z}` YTS7ħ.Din@~uu?~J猐e;WڞРZJZK1_S\RT*m_@㦗2*k69Ŭ8giZcJʁjyZ܋rj}z=Z;R쯬;R,#E_D1ʄm>}>9dOORHuTm!ڴ{@K \N2 &y17 F' GfgLz⏻דR(`o}zò~keˍGй9M|*5htdD,>Zy(3(zuZ>=љ<]^\P3Kl/<HgswwDӫ+=A LiA8XշL#.2ڐfAKŒ',NKMItk+#@+Y!hZ7/'xH,9,$g44 PI[!C(6JXE ƅ/kpd ?;Ff-3$h碭U : ZG R*RI$ٌ;AyK?w7ZlnoӛlPV,N*$(W@%j~i)"sZ3!%1I&^\hafF*mͥ![:ZBG!9]O/Y+I #Dj <⬯5Q,Z}Ta#H8LǠZjL͚z.G)( 1%9U vP`"8Hv*/nuYЯ)3i>K>3tLfSQʯ<}򬾿!BzÌs@ィڤYs$y>\,{EuՇ߫߾>|_O9Y>Qr/ ݔa'R%~$H\tv:BK.~(\yY)B%!VW\G㘫Y]mF(rFjQLF@B'ʜ9V#P鉊x[c:s'mX t |WÞy {F IP\KA5NbCE>G<>JB74$mhg6_>{-B׵ %<#Abԁ/8s*Xke A8%?u ~qwon9ޓ )|'/{]|AC)*u5%t*~fr>Af5JH>1c* %KY{Qtz >Gލ/ECAw2El~;4N*Ї6 "A0cRtQcɻ|ecتN"#<8}^i /{1RYg;?Xi{E/\ 5]pؕE3rUԀ; ۏ¼"/~ y"LE4?tR|5A#-j%owfֽFpVpGyhJO7Չ.oY>xb9EbArp̃М}gڎݽzѠd6l2 5OuS.Fz  Dj3AAz.i [쟍vk)lĐa%OvPfe}ί %H]L L+$fM8Χ(]ItͼU!ǾtɤRd䈡5ͤHjC(7aiYr7BB`"q I\QUzzel3 #k5 Z>,3K3=nBz'; 2!J <چ5w8)q&h?)á ƝC`T7  ?Ƅc7?Y).NVIA7eS&sGSs 2J*-2h܈X[㔏M,!BZp+²( R+&eGB욛ҁjnJ[u7FįfhϜRFf" Y& IFtQSʢ>JÈXnyD9DI(z@ *nK+m)3ګzbQ}]K]H^hZh!E/fH1ukQFYGcQDԊ95g13LVE e $xm IY(Z2 5e%.1}ைZ/0"މs+j**ŘRL] z2&2ggn2F`@t>|FovWM*8rxr|N#a [N  ]t2w&}Ca̍$Zt0Ilim:t~.{$WIį_E]U[$3s`"Ze ĚPU+b"AS)7DEWG)gTTV:xG:tIHj@6)+en7Oe6%cdF"pРq8?#a>hu/6(NhZ'T:Y7tbJmsW~J2Hin["l˿:KMދty@1L:ENXAZ8c/H}|&ӵʰ>\?͂_ ;RYd(.znL2Ygtzmx&\L_q>0?ʥ7{׏7M#iʐ?gvd .ħ>GeJ)x97GYlLV&Yɴx3Qe(炿~C BJCUd%ʢ[tAjpںj=mAZ0τӡly0?}wyc(|?+9F(Ya&*UViZy-deJR Q`6"[XcY\c?zs3hBH,T Dq VlJ6t'DC DIC ', %L+`0F%ou yfCSfCvgʉPRю4KXjڭ]FLXBBDٛo$,=i%JZ⪭%)K\kbMTNQ$2,О$Pc#1jkAh31;et;ʐ'FcCrh' joxt p |7 N2ﬠLR: 惰4@d?GK%ѮXe`˙@b6py*h4*&JeS`42j0B#oAWZFX4~,3Q x;8-aL %O+Y囲X3SNmN{/8TEޖĚX:Qv:$RjekB@0m^@,IՊ%6oG,)Q;=&N@6+WۯrxrxPj+C [Vc~@uZع~ ڈ &MȻ1e! π(ѮbwT1!~)hبeaQ<[ݜ޵ҌQ'1jdr:f=9%ۛո8ހvއEV~RkQa>WkQJ/..oiGkڷXﲪh$%&0-f& {;dCSF%+=CtF$xt9ěOhbqRVx_ 0uĎe*<{qq-߾W6f;xq܍ҭ6[ N#A$1ܝ{Ɖ "y>B9 (~ abABN*Y# xt"# k,r̊\U,h-H= z*<"Ǜ)(*sS(!B*ӆO") e<]tYzyr£D p&HBI" S=ݭ()7wqI֠7wAow=4 [ W%:|0F$mq&X=鑶oz$=zcEF+7=rԺ4X \,&5;M9F1q.krmnI-)€p/ cTչQvZӫ)8ӯF( 4aߴxd]T% gdrr`,wX2(Q#-~h J䩟#8r bpK;);?q+ڢ u?ahH%JVi7 jՔ<|ZJdF |V-ך4U$xE;9[=`Ź+C>Ko2)32Bc+knJp=Iis-ZEc7ข IX7iuǜmuqWwo*ۃO+F6Sx%zR;.΍֨Jt@"wl[aP Ԇ+1 ƯrLMԘUW'^U]a$ÂCx?ɱŸyJ7:=*Ow_;2^F2kHN5u7"ݣoA˾]*؍DBQ hJ>BW"5HVNB] o 8ʊSkežn=mWuVބ[[ k#Ɓ}z+ѽFyk=>h;6>L;ӕa]z,ZnoϋD4m>X!8t*|SOn/P`i%1;Tz̧[ILx3{1̭z LӷUk‘Eu!' /3ZW;܌{Lҍ5! C1CfmI F+ G)@ ]jIqXykFvd[k cЭCXS%a(ߔ%ߣ1 7JqaIWNL8VK踭{*Y~t4s2a~gڍL 8AWo=W)KĤ=[1g$I #VVm|+i=c`P T[z]uk]ceS/(ć /s6Ύ`#]V$O*8vu90)kVH!wjV5oIz;8co&-st 93g1lcJDp;kYKnl5X=Zkus< T=~"/j1:~6qnYN_}}[12K"83JOg4j;h 8eG"0:AJz*Df:e&"DXꕝm g yזW{wa0pǵ{~Ḣge_nE?5e~u8zr/g5+;DK}tfwOߞ?^Yn]~ -Yqof[^U/_Ҟ݇!ޡ]2c~WγM@wxj⼖vbM(df9/V"X)GCs4>W lzCiJp%2X2ᨤL(+ c{*HY2aAӡPٟ|~^~)q9t-|Y7/x3rĦ@1 ˷OwB奿]?^ͮ/?Y~I8#mU#V[1P ܗJ,b FK_9ghJ Zr./X2&FiW5F¦`K1S`3DAJMcQer H!S$CCج&lZ ۬Y5*6uȌ A9mcy)@<u% &† RJ+7Mؤ,JrP .3 &."7wzu `C!(MTļ"\Т S:HP,*f]\[YxuXfZ#51+^y,X C%0#9FXj-C{Q ]2U"]}R)&jw[sw AHmvJZѓȃ>

R0n$- Ֆ UIL, Qv(Ns ! /nXl"UϿPç?~j)S>[G/|֊ k0\ϱ$kr72ݗʏÇj^?-Q&(KrSF-!u&3P/$Y'1=e ؋j aZ/呢V叫o8߱-צ4RHsz/wE(A7(;]`d+Ssŏ4Po&vO!0zY v{˽Ec߲r˥oey8r5^B~FZP&ұf #sZMa:|QBn ei'Cztb e+݀rpu扟l'k4RX,zk{]IdkޞG{^ YjT,\ޙ9U5plӽ3p?v9 Lї*ZP' 6_Ƣ  M쬍Y%JXLOf?Ƭr%!g^n{N5ie; k~܆jcIXu1ƴYqjcFbgp"T[s-e"r!b\X>FrwZq*V4´0zF (Q1l$,qQ"fDl%*s1@1Bnݱ.kdgNkZ\a?_q_.͞]Ƴ4Sp@5gѽ/$U@,khu'hգ mOVsF؛Zou/>nhWߔ{aY |U;ʏvQ1EC2̕H OJa_ݶ>؟T bX'uR=M+mkHROVN/ܑ?RkPqir+Qt5wI 0rFM%3%=)3wiC9&';-Ab:﨣ig13|)R5!ϜExJP z T8,>]dBik*]<6An2I:{H:U-4/&zlUp]s0!f`?n^}VW~|F j'!avڪ┷ڼ;>9p @Y1ySwp `^ i0C ݇K,\h\[iu'i}Tj*5>N>CJ QtO"FYQs{"&H"Heg4t;=s~$UBJ?K=XoO&R9Ip^I98g7I $|eWvYT]GN⢽O_F)0{jgӱe\./3T y&A_PBvFTydQ5 x̴ʏ3#~u)q9vAԩEEzA2WEC]C.֥X,+; F.( mLA(ls>npϯ*7Z8|>{o Uc__ya6_,aw \F0 d޺I;zKMOԔ:(]@P M1MIΘ8PVca6,qԁ68Ƿ ?7i:. Aaﲹx%3>.?I g7kN<,-%wDu\G)dR鞅"v*݋ + CYԏDtHGL,rl ð-lg1[@8R#42g9G1V BSE[f4-z1*.<)pR$crĘPP r#L,#bGLJ^88× Aj2*əULLcIn=kȵχHXϓQ T xJ_N'С0JX?!s3>1 nu0-8y hoCrNC SP,X&5c!VFRJ1P:A֌a5pT$9AW^A5悂M1v !8+1T^; aro|׷C̈́1͜BC>E[)$ IׂWr/EFWЙ2 e;B$R<;N :2 $a; )RΠSU*7Hyh?%v2Fd# 4T!xGޑ< rEXz2.s1Z3B6ʉb OH..y).&Ar.FF)1D lB^f8 G!'*+Q M@D)f 1y)^=h!,Ǽ?5ȽUτ>\gRy/HRձY (aD 3ȝALQ`:jcbW҇%2s@>ʪ!pIgHĽ+"9$ }| DVjݠ1 'ZpBW;lwLTZ@ #Lf[2ɸفwlf󶘴6TtxvJYP|Rq$}ZXBpt^Dd"AV$FVÚ)x!I ID"%'WTrB I p^>}RBXG2~_0# ^liġʎO&fP>Q?;96]ۺsmģRBz3U|9>Er~ QDWt6wNwO!N[{MX$w?ȹBar(qaո{krw ] %ʞ0vg%ZTk,n˰gr[fu6tDE<3mo瑱_$s, Kg6[hii yt6hwFCzm@NF pP[U3V[-x@=JeI`c,T?u`uX]cViۑչREs@*aqr;cDZa)4DSG= Ef"P|GR# N輔1 9g̋s$w*  amC5Eeކ8䙳hOURjT3*!V0/EB/lBʻ;'a4TOV"dm3zi5!NM@Mp`$8prcp `Ք>rrw} s%\AʞͯE 'Qx#pW˛W/{)@ϑBtsMI"խu9M—2_|{]0uvIad |c.)QeS7+TTlw#>1 aǧЫ4(ΝX7Q7w4&IG(Q (koqrTo\A.T&`tbrͭ!.gTy)C&MB3("㸏( C1CdRH?5F. w!Zp{%Ihq{I@`){3j.dXS1I}G2iNOSyg%K@dRY40Mn^ I uQKO 4ݕ&Jg΢[<#y ʆ3%ʼy¡H2Оhma?nǎ]c( )i#_E;axyg~م<:FUMI9deY̞yW_$ɍ8`֏5cWk٘9y͜wԟѮDՎyIuV`Ve>Ψ6 (\ Do~;a:|?_˓q_[5 XAQȊ LmeP>n<qO8*+Qy ĘPj(vc4/o0xR"qo_Y41Ya3 Ʋ‚C fi 쀶Fі/s4QLc=Hl3L&>~2{1d;+yssCY3wˣ:]k\劉|1敏@~&h0[<Gm _P) p>hg|7n>;x~f\?Tg&+vws{wHGB<}@$d]Tx뫟㮇R]\TYdL立.[R3"~BN0͖ͦ<|v# z=8N߽jdc-cϾ;oόi3/csv3?0:_tc);p3ŏ\$7IDPG.:gaβ#}7.\=U,Ԟ*ೝTΖ׍pM)-漏uK FuRƺcۺo$EZ.4䅫hN'ZY7n6LFCilah5 TC7 2I Isf.~R榔 t8H>iR^XF,a0VD/Ӥ$oX-56xKu)5pT b\;ay)a]h WBɡ-&0uK FuRƺg&uK?iݺАI:fՆٚ Ô5 E0 W ZSa5 $DGJ# ŦԄQ8OA0M|G+O{ԎF{4䅫hNa8xe( NhR1Qwn}Ym FZ.4䅫NHǗKc^ޏml2Z|폋tiq—+8Tr]9C+CM6EԵW߽y{_OߎGoۘ'];Bu<'ty Qg|:6I|mN+t9l*@Z<Baǝ{\cI%wJ9ѹ47Z*xß|;O$v-;-S8ZA?8@:Gu@T"&+ \;'8 @bXPH&n I\%4T5 (ZԨ1rq trFExw*IE9u8@m +gk?yeܐldsq H5qQ?e;nƳHDٹ9\P.yg HVS}<@f:]Eo_h~ c.k]|ɴLLt]ơ!% 13@|<:y"ВbF|e$ވD\B@m؛|*y\Jix_J- kCDjreuZS7_Qx"9,Wyx(ʲu[*bT7Ʋ3zsftٺ-FyFL❧)-L:4/+8"&dY!GH:eB08sxCG <%tlt?Ӫlc>w)![)c9=wIOc_DvJspOKvHY>sM.ȣ !e"I]7ԟ)B?gğOgE۝yۏ<^MA ?g/2C\;S_.`L&QRHiz!7Q6 SG w܁kLt?} K[zQ|<]y$B ԕm "Nĸ[թ׀{cbt#J7[Pv ˓~c}<#X"-N{LB0%;{"lp L[O 08ҙː}?# 3%v'3; ac`.Av 6Ӛի_Ŀ'_n\y\ޒ(Z7zru\IZYoVqW? R+h qbKs~İYd!3߽Lc78>qsS~??e l lTh2h+y@w!yg! C&\m7~HV#Y+5}Tnv*=QU&:=(Jd$%Lكb'%?MmT+$Xxa  W?=k ۚ睃='"}?>: "qnOvoqx;qsWcߚx6x٩/ŌZ{Flȉ1 K2s#:U:: ?_-4x4K>}RNz8S+WKIk@][bPbl5HS*=4r!*\˳|\'i\`\;-pwgtd^uJ8G&~p_.6B};~W-녖@׏أT7rn`HqQR 1P(B4&fơOBOTK ab!JEXpTtРEh1R+& F Rs6]롔BzQ"1=$\Xjuk Fⷮm7 HaG=AhH))ʖ=nn~*s4V۾CDrYzr/YZ ĘP0]\fz\&.dZ%j\f)F'eZ3̕310ٮ8Aץ,@;?xNoq5IB#Ɗ }{#,2% щCb |KvʅZɺ敯tV5mc*B RVQ 5E@V8Kn 䘇G%zVyfuc-~ƻNX<'VZyODʱY}d5.(,n\\J6o jk]q{08X:3B?߶c;N2& dҫv6*_Q(('ض{@f?_2ޗG1'_42bMNĀ y#'B Ht.I9%,NP 3|<@lCR:r#EgU7TPX<5J[7< tQ us?ybz)N$ͣ8BDNAA Լm!@uLI}aҀ|\~-n}v{qB60.8 C.>Ջ躺eښ۸_aek.%qwRuv78N^RcqM:I忟Ɛ䌄!FTlAFhxLZC<<{ Kar[G W(/`u#}-~pmt;XJ[c=({wõd(=_XcTRJ!~y `ɡX_8k_MÏJQr(I5he^pҴ8ʐvU>e $ q?&U VjNiE0(T)tT_FF>\$NG_Er;kuuTޡV1Rm3Т1L.2t-2Yw8Q\X;d"NELh"I&lBDj DA BKc4%~rfn>G1&7F{Q@oFMm/Q#E>hqe ~,F4xP¤1$TRA` R}DB{uSP%)xĤ _Vs\$ѐ&1D*N&)&;Lj>$(5.#R JXHD6QFLj~+A(61 EA) (b)'i(yJ!6-Cl6y*))vJ9}BH1ujdԑ nڊ7TP;)rCVu6EfRX2R\& PY e㞛(< j:Y$$#1JL+Ȉę䑶4 )5E2ccjR @UYp؀ (dn#]u~E {Mr1ϫ4Hw>gX==JV*Ϲ_=ru[]<|~Rx.[}orL`\.ޞ-u~Xo')yW4[|;]gB|XFbBoF%DnɆɓ;q,yZSCiHV 7B|Bv(iU);ّD)q,{):"@vddgTH ၗ#50<(c0^%ǁjKDz* jKj!Mgw]zsekXH IWT+n84tq< SPzu#X"ĤTkC;&ZHD y\}_ۮx|o/[9߶^סHHݏQ=voG8d>~VDZ|#{_2go΂gwHJ0ع@9V\Z\{~zTJ:X.^殪U v4'pvh?삭A,uc6o..j9Cb\ 8uϿ$ Y.詞&֢@ހ4QiDR-Pnţ#W^ ^5}Ħ̈8dB.2=!L1gҰ1Ji1EoƻVB@>!-(rc4M){b:{1m69.0᪫'M; ЅJ^bD;2{kiHAr).wXHS4l~{k(Fqi:fk9ib_iik>-O~gI|+|p{ +4G2`.?Ma^%z[%ZL1J< 6N1xEXHdx5&"("N ((M'tEZ%j)wQ!4]ov$W̅cu#*#'uJ',$e Q@];I5k:;&xSꁧɣ..ZٯvMuӻGt3fMh`g#fr7EDZy3'?%Adl NY0 &>v"'e/#IIQ *qg7s$\F w"ܓ$I@$ $(|f B59S(%iGy<><@7#獒$ FOe2=kyTRjUq)/+uqzzo]s A~.qnbTnvрx4k_=nOzmP_<,sVGר4Qat TYBS\\"4 e:=t|{sf= 9 rP o.}BrvCnjE"٧ܻj8.1ϮG3fO;x~2M?ŷ.jڻNftt3[I>7X^\h1 q/Og.mrLȇi!U͹֔]݇g뭈9s)Ot -'[(+wW.A2%ڍ*Uh ңv GtBǨs#Ou!mݚW.A2UH9NCuQ'>JКC 2Ԁ='4Kev?ayk ]d$-eDj[ˁWN.tV4Aݒ sn .#c?6{E!x}I!boLZ iu,|%tT*kKHր>nUL ŤVЪIi1& }JVx@u+AHvWz٦!8':CLm woH,,ys5 %wFZ>9/ޗU K6P@KJZ,RҁNI/L[R+5}d΃mJ6/T%͞HL9^کVi]S? 1 zp^XW%dE.uj&YW^wZyio}{!u0Zx<8ߝ@8_ 5cVY58(Ѳ|6;g`9M İfy]v8kH@w:Fu]s @sK-vvshqO]+P1+#^*&]ZVGPCz>Pz6 aO^+9W `6j:"&J3eh%届ȬWAfylA0FJSɲl'rMb#%R̤Qdkl҈  bXHS(D? X Ž?e~סTig7H)H(kQa Pɨ{FQ-U VBl@ﰒ4߯>! ouFɂ2ҞLC L'm * 6Hb%/ӫZJ%8FFNZ}XDS7.+y|zVP$s]he嵽5QtOwwO7x]fBr̨1w'>PȎngPhtuLڔM7%!90ϔ,U‰/t OI#򣟙0HA[zC>ևx`Xx@ @gE^0KxȊ|SSP,<#@ ӏJUsssPP=32i@8\A2-,>z$SoTstј7!r{xR UKfȠ{ݔhET| ljU$(1oC F+=9ɳ9%389;yU2FMHwb5碑(M'W给Ϧo(:.rkdd q)>VG8I̘Lʊ10UF1c<"P}(oŠ~".Ϳf/KO}|$bbAd,!%côkf,Y1) lՋm˽Qbp~Zz7/m Ɵm$²mynqbp< %aQg@oC<*RV1-;M1ͭuU"'ѶzjY3!^QZ CƪH٦t+Wba=fXcӔD`)G3!"Dij#>Q "IB#9  NCJe$%G_Erw/77dkBevv4V薑ܴP_`EG#!q .Ro-S$b۪,nPR0$T[lB v80_0QPK)9*K9y29܀@}|z<sT_u는{dl{dgWH켨APvP;7pb:7ڊmHt?o'Ϸ675KW.~|-Hzv-&'9w =ݯh]!1Fu9r+8gwIm$DH/D ҟ۰Kն^'.m1ȆǪ]>b&Xn,S1 -㙕pV[ЮCjCaˏm\A[kn= ׿%|s@f rHD%hؼڙZ+R/lvdr&P8ILYks@dñ5ܺcʒO?& tYCCO8 q mEN'w ~x+.k1~klPBjc AD9ڗgNU5:WR-*OJAiL)DU(unє*/Krʃ0ι%7ZJ"_ `4 |ưZx\WׇOW1uZzyh7?w5OOG Pz?< q Yȸ?Yy5Ļr~һOg4M1׌L~٤%'~x}> j%p]DVzZE{tIki~?ku;"TsJFFZn !ٞM &`p r@ }Ƈ1 %&g?>OȗIf?QF~#JYG[\npCĹc{dJI`ߴX3汷]ɉgN, )bqyzb,4cH|AT )al+VڑnZ' y2=Δ#8DS!r܌\`d#& N[Y2?~/kG%0F} Y)ګS97|t]䆵Pɂlc+ˍ OJ_M׼"5+-!32V ˤ*yRl}4⯅SFNJd7'~pXc'qa[7_p[/os(SU8T6H sf>ުrO -R `l_"+Ǎf~$͒r$ YrʌYstn:X 3i|YH9MmtjII,y8(ϲ[Ft:T\w~%gJ~cI,C}]>+66xኦp!-Ԭ+ݧ LRQĈ"Œ5|z$8E^ab'V^_MvSk+gu+z[KkcO*ƨ6{/sYEuNK,P)n5oA[4.}=xfeBfEf0Bk< kRNbMr'1`۲3fɛn?-̊, !o@: W|1@DS(?#!W233HAsSKKٗ?ea]o֍9y롎NDI Ҏ-RYT,o (J`͈}0iF6ߔ2h"F# 0+Xďqyw_1"U9>J2S;l䎼͕x$p7kjΒɁys+ENś<S\I%ėF7S>9"?ĤO2)Bx#W1Bx#Wr 3nW ./lN۠˜KέBHTlkOYt}$VObRIF2բzdqj/z2-Z|M>0QeZeMWDzDoEJ mS@RZmkiWFUvo]oյĂP%8!L; OlȴbSbl^k/jھ|g)CN_nbDd;]E*NWvNu.\AaGF:I#}S(,7Amt:܋R*tEC$ʯ8g\[a|3. ?=0x7 G?Ooe'zqEwVP}䗟ߝ8/L3gESM>_|7ίӿsDLJ;z>m`$)@$!e=:}A$c1W!8 ?|==vDHY&59nsTczAA7DH&ܠJ))S9SuNrAa&Y(Eh>_■oTM({?` X!hk 6XRi.A9QK "dr@2R yФڸOX,ᔁk0C C2dݛ~Q{,Ig4# rs1zߴ/:^?6s%c'pBQ8Ќv_E0 QB s21zw7!@mcP>VGub5 Q\aFLq7aD^Ey%.Ғ<‰iDŽ2׀ BpA{m3A8ti@ d*-kҁA֒*}8rBzrV^&隳/įpPAjE.sZc(29t"HK2!E0hzd}.Y+TFy Co+EBwp\U C)_THcc(g*+Fl[>u1wO@"WP7o̻*ugkdD29 .L,rˠX0W,BŊ!E2)+gtIzTg?=e[#ϬZw%Q?9Ug҇y OHς0}b9 w/&O.$c2t8ksĜ=nHm('‘Y&`K*oUM ]={2qY#نňLmi=tO>6& GRQsPiv7{Z/W,Kp\=8ʘQ1gV\+ ?0TJP[WTApd箻ﰃ'`~IihLY`0k!x"R~ni ^N9H:|fBn´fj7 Yj \q1+bσ qPԫƥUӔ HoQ<1| 4c(sO@ ͏PVXlx:;7P+Q+RI ouຄrju%fA(Xh9sLI,dX3EBY ?R[O3LFyQ/r)^(_9Hˡ@&!NbHqI!H(#P%D-9CJOEL^tKNQպ(:dA%J./H1ƧE7~wcppwX|h`5\˖p4PEbTd` RjoR>yL绱S/ɽ8?CK>3-&bw/@4a"s ˾I" Ct"/eqB1c3%JO3|%pLBp jL<Qhʣ(B<{i0YrLx?1dO >1%1Y9@ +Τ΄0m+2;ˉ5b@IÓ9㔕L-wCdHh9g 9eJ(nA%7w.:bXEo,iS<gX[)AzIJrz1䉪x!a1OEnI( g`i3D]C""1 5To!%|LtOLs )q=X5_KjQV1Ƙ>Zc\b. ͮbN]s2B`&ˢ=%̯^ݛ#Ҙ9ԋv^$Tήi]%Ng `3W*[ Pp+ѶGEޮGTZGiuԇ‚(w= HzQ13%+G A.Tif:X9 Ue %hN$YeVA! %//INu-^/T1dť4k%A@h8ŮFB a#51\ )!)bN űU lU{rIGyAB@Z[rajD*wm9eB,){@j$!#3juѼ[)-o6V<:vkb{d 0ͥ[gﬡ$) 1 @t =E${]nB'%: |T8(^A| <ئPӄ{l wv\T.;q猹g0rHmT|z0R;zҋ)XdhGI.bE 9$FCpVe'힪[+~6NJ=ťGch&^Q(exO^+OY| T~)܋DTX_20,l3I1$8 EGPpx($"=@(c\{L`LD眮%qåm4!9`+V:tO5yX8n[iBɡ9$ 'OZ[~3BTAf9%c O[ gZQG|Zyq@5\4b XmtM MV@H3sJz^I@ň')L vy1La*x);*&YC^}@x)g*C)`~sD3\A#`XMB0x&)h?$0y #(^IgL đ4v${SBz-( cA%Ø? K,$gIp}AXj" 6yc)ArȀ/]H^+Yl/;Sy üñR^ ñƭ y˘ Μ΁SGg)OIxIIK0`<aus>n_NQ׏<Ž2`L /w _x܋@cCveS \ %AZC\@ZO>@yZO@ _:CIx `uL\Qh xs); =8S\>8SE457<'h'4;lWl(M9H&% =PiCeo2dz *Ӛsj#1D(5P A9)F>pYKIK,ŝY{D^Qn_6cAkN7*8 { 9TF5j!e^fo7ZMO_{+?'.o|XP.矾>[ԗN^z;>,iJ7cځo&<0bXH? l划jc>|=LOt !8&W'E'+!k8i^^bW p +VZ➯V ZFnoWB\ݢVW ޾>1Na߫7֭ݬz eLKWo13 *tnWn|\ofh2$0sVm.M4kaswJ5_3۫tO-lY) ire<h OV\Jh-It+Дb-y-|,S K7)[*1:F6b-y-|,SdhYKiz p4hzR^rϕhp @+3\7ۣYS: 5&( 8q^Fd;TrFZrOC>sM)^'>kLfW>|֒)fۣYK1$m:Fd>e$tIC>sM)f/7tK uR#ݦ싏 y~{H_<8*.KƤ::ƸRchiq͏>sxE=U>.2_-XQ{z=+j%J)Q WE<3^áʢtnI Gx6*EEڪ5_LxFǸPlrZ}(0$kW 4r]IdN{X㈩uߴ8!CMJ>&!Zij (NG;&LPl~>* I x' TqQPP_jvRTz+ 7K'PׄZJ(3BKVQsC`Ky9=j H7J=g1Yc^~RTpNk,<6Δxlb-864З[#Ï)^,r~g"fkBB8 L[wljkB^@m|RY/!,(`>0;x) +Hk^x$>kYW,Z .H GL[wa<{}{4k*fAW|}7&ER2X.OKzմhOa4Xmτ-ItPqlncFJ`֫AA`1,r1J{lqss%O#17UX1}<{`#1vqn>~@sLGfLįYc!1mոchϹ-<ЃXpm]OO#WGބAx R`'j& V2A PP jhR$6A^0&{> ,~T9SG}E4\7Wu#oehߒ<^𥿭`%.iqm1(t+?ܓVٻ'm$vko콺ͣܗ@QƒFCI E'Sx45n#P!7@f-v ށ8CԈւPnLZ#ȑ$z`ɏSHglYLj4g uwY1]rԀj1ZAж*#Jj^ZaC㐇UЅyݴgHVUQ[g;Ŵ{nA_ ͐CI3;q37g`3;Ɂ1$1 ,j=Rvɐӆ" =plW㼷;=rftPݣ\3M^uOuc=Op8A(pǸ# 2yE'.VXԐ vb%|Xɇ?y!8! g_йHi81%Lc M |Kee%ɂ\*wL課gS-Ln=pKLa撕9a4|ٓ |+vk d)YH*e#$K4lFUҧ*A;}EAlV._1jtg 11CǕCfe|/}C:!Z}ܙtH:lCRzlm&BYRm9u8ѻ"`6552ÂM>dV] JR9-DF()U |)&O@SOݻ"gzwE-g(l ]!J5=Hwr ?{0F&?ҩ !`Rdj"2?|1r,$ՄH[ۛb&]1~z}f>u/!Lnƺ fzU>@2ZlYT0rRLeYJ5{^<ێ[cvUYϿJa809asIX&ºB](dd(ѹ4(BD:l=|5bCvVW:/ +r\.KR> {zQr`}߬oOa䅻20Z F\-)+u3yVX[,D,SXꬴ5HQ#%cOX"򪪊E8CwDAlkywDINF}ӵ:VOzjܩwQjqƭ7g?%-q-Ob-au헫kK#Bs1ZϫU[r!OcO&QT*BT*5Pc*x[hXxDɌ8N3S/,́ӇažkVt\9 K$4LG|+(w`+Ps>1Wp թ tJ]X)3_&q2sD၀ڲWL!:c``,3 ŻܔEEr# S|)& FmAx Ve,Xf(895%'F<\f Ge;[9_. ҟ䫓}JRP2Х߀6S6åҕ0"o;Rp˂! 4\J["y"R%A3eIp faIaY2'7Ĉ l_m=AA 9HAS}$SG1wr!d񄘕{ySd +K(r g*s8 .s媤9Gܔxp\xNk)Sk)Ҹŗ 䛵'/M4I.RĊ) Ǹ68 8)PX^z.UK}o[-l6YWdu&;s|lXWQ٦ɏS!{ܾs [dn拼m쟥/+U"#Ay(?ܽ %,)>Sq=9x GnuCqmC̺_9m ڛhwa~pkuL6?R"Of25`QA6b, óٝΰ傘-7Ӌn0u֭9AG(T.uRpU+1S1|9;o(a#@< ~zxD}pXr6<<ƞrfN)qL[ ~x)7m.rLxp. tWhǽ=Y:ryxR=O6E(#9]]'"u-HiSMq)tnA ")+)XA )B!(! SXl ֈr] siLY }K< %2/pHRH[pp9_n4Cb< -W:W٫^orB0 P0CiRF)ĵpY[WΔ(KN8R -z>5wj'qUKUK@|謙Jϐ䳘NlYr)Ə-%U. 2oռ,'-a9UDž'`-/es}E Hb4 WDo@G>u<%(ur42lqPn3Jj d2C4&"TbP-I* œw JdԷ 2DaL&2Pt:7րd,7 f-9KNk`fRcVB\IcM=GRlv*^4 B"9,8tJ>34*d%2SL+[rR2e<,lE80TMc%y q:%BcJ1}0\)bW8:_@"L1}Ͷ:p-,mviZ{X3q[) Kr7DVH*@c`80!%5u0+D8U(2a?*,Z 2F-,PNKD zgQ+rc~& Nan=FÃwc#c[Mkϴ/3ps矷s맇|d (ĩի:=ttNj@!ąUzEL\W.,X]zE}]1=@.KLJxFX)b@)(+B!^")SKn ooΗzٽn`@+%q jSSd"$˲%z!ߪٷaطkct93Jl: ,-)Cc-zøLΰcƶow9wf]-1 ɶZ;w/_j4i[όu_y}|?]}5L߬V*6\bH78X| 欖}^a(?d@FD9Zu6l58b91$ ̨'PZgjz!%S)HN )yI\vK 0DE3o*۸Cpm$ΩZ(8W*[Vfp_>8tӃ3/%gᒞ3;gR/*ˢܱGph*fw7 )]A)F"AN}zJ>/o@90̯o1J MV9…UIǺV*y)-hcazc7Ȥ<NˏZH=<.+Tr]U !"JbHn>;Ƅn Nv92f* /*BsS rBNT2(e7@R#B1j3.󩦔5쭿oZ~c>դ.fH!>rq9)&sSx}ŵnIZb|w'[ZPٵafϞ利׫>}ǣenx6 *]WDnkq#wۃd2hA"WB=@l7-/ /ш58g Ƙ@1-?%+/WXj>< PpRBYoeCn wb(FqNjnp8gz'A uvD]}i@R*coKWX?=8HU GlVn6><83H-݁ տ.^}~U6i/#6>mޭq~2qYo'aoʞGAţp)w^Ђ2W8NJii_ |״h % 廒?^ :Qka~pd{ϗ?|߼Xsor%)VW7c_]]L] Ot@Xۜ2|sňdEHt}#>1[R?Mf|2+{kg?Gu" 5?.tg^zl>6^7ta_$ɻ8n%W{ e}سH} R,K4J_rfd$v;,k8dbX*ZOHIz+ 2mL=:>ȸNx'gӒoȗ7)d]~BU :M?JՃf$K9H/k%V?l~ZS'оG7 !Rj T 5Z亪M K`~vV*PWkP4 ADiRf)mUxN=![)TLOA=܍"^,nX aQAN3I(feN k*IF@x  5@-pj2NEŵ?՚ړjmtR0/JۘTy >0:"~IdL *J#dejla|RnjAX=rsKiSʂ\PXzn EJZ 纻jvd[-%S;FvD%"ڭzm ELiM[;}ROG.fNߙ40\g違Hְj r(I^/\K!~s~6* JicvyJ&nCqI圱4skt5zƲcyF@*2t?Z1JƬ_m.g}9JG,_5 qua5>bQC|#Q ǔ1WHqwWzҝՍ3Wkߙfy+XrJ,k A2FUY&_Q2\WUZ+PSCzdڇhwYÍym +vz9-çicrM./>n/ \YG5>}nl8]F|;V;q22g.cۣr󑈣Nm; #wN B tiTUwʽd օ9/yJr3=Q{SCL;~_V#W뻫`ٺIޚ1p"7(lp~)Y\==ܝ_]w^z#rK$I XTR&_ o feo?7fPa&_/aQ#P:%uŠgPMfiѽN_WR,?j9[Q! :mK!hn195גo/YrN+#/e҂lF«jO8E8+INN>+!\LXMB1 Dld]ui[#qr>WϡGW OS2Lu%) dVm٣xi-uZּ5A%0hOwOwt})R|o_Dل+ۗUəo_$|h' &kGTϮP`mTG(Z}/UG@81F'vj%@LttJǭ6yA$ 2ݩ Hh%tJeF".O$DG H(.xLd< FFcү3P-qHتQQUצ[WJ}R6 Q[j>K b/ij k4 9H=V',-e@v[mt p )?N &.駯ӆj0d=nyKi# *k LFU * cJAnQY}_  ޫĤ4g YD`*D5QL6ҨQ%Uj4+ĈK *dUVT{8T2EɐLӰ*)loJ;fŠ1IQAuvsW0R_Z,j^mLs/-K3\31ARvDz|;ܓ_2G5Ypg qɨ!0]79r_bH6>txǘJ=k"_hjE :;E 7>؁uxM  Ef) |xyȈQ6bQ\Hƙ `ut7DvgoC8|gz^2AU=79Cs.ﶩJ͢(}-%5qkRWZՓm/,<|"Z%S^h7%ZcjJ DrNxWZVDڭvm EJ8n'VP R5'AbMPpGSW7|bМ/;,<[zɣdUp/d2' M:#Ss6j\YE/OQvFb97Bz2M`$DAX T|fU;H0Lf!D9:Fbh5ECTXi`. _"\Ͽ,҆zBwWhjQ_`ӷzJ$e~O7s}XIC-۷uSӤRk~|qwrVHV&X2*+sz?7KMxw}yި#pDd%g0yJ}UVkG [OپA?4 g=>t`$q譌*PdDBmvD !2CK@'u@~wBܟj1^Ob|q̔ihvzL)v+ݧNO7U2 pB t /3ewՃ-O-֑-e:Ȗ-.ZOd щZ*nt25[T:y%TU]?BqIf ɗ[&_1 :OQ.\&+Fu$Y8"QVߜcWby '7ۇLv^,o)3orUj:YۓK:s_ޅp %m ؊}~XNѨ_~M쉞 syZm= HIqi1`|m6j5?ʖL95[gqǨ>iq8 DbTx!q¡Z$=O!Ji'$&"ʒdr >ʡA᝟2mUKa]~T`C(jɯzp+q:kSHI" L]]-5TS*_aaJ~hBFۦZVɵ#zj1(q[Ԏųv%̳Vzֆo\Dd!x{J6!%}BB˚ov.273 6#|19x0 m[_1~Em\ rN0:]xs]޼j=9#1HxsGRsZ%k"dЗFLuFAחv/ j{RC{4p܈z]+B# 35frZךM򳤜I͕DFi#uT5;&*HnO"N9f\}!J,1Vo.hԜF҇P+n ËŖ7V"V5IK{S,)߮"nu6T"́1%}ew!|4ȯnޗ(qp*9=\ YÉl48KLxnf(exl} /Z֧ƇNyA2\F313$u{7y=ogƛgXmtyq5{jEJ;o W'g~2wh·U0#ٟݼJ&&Ruⴞ#tNݬo;SMsqd&}E.P m @!O5P0[ΉqҸ0;.nX!Y$SUCr㌫\K~ Ľt8y+A|͔:>/^\B}wsg&bާQ(7WدP7u3dAfj !|+n zOɲE Ҡ%RsI#80DR"D`VS*!sbHO1YĚތ`HwAevswA%= *xژ (Fg/z{}F8GVcalZG,b 1 \j3np E 5 3p,KE 3G^$~? 9;9ic!d=G%8cdl//ky;oh_=ݢ0Ω靮o;ct Ẽ.vkgnUv5>&`.SnogdFy1 'F_ZOd uRQ[嗎eUO_X^ӎn8(5m@=8ڵd|E\-e51WQ)9~ =gCK2Cu: pQèBp}XbY6b]!PW^qڸ/keK1SSU:jb˜0KXB:( eBFj&(#g誾Ūj?Ւ"mێ\+WC(TJg~pC{IR&OB:f`RD;ln#_?@M/U]]ď } z=c9L~lSˑlDvLu]$sebDJ;."vDt$1 cb4[i AӔo䷛!|fw|f(D`H9jD⃋ {s)J%-S)t0IW8sƬ[Li'SYVzn1x iDpS<$? .(K d4H0#x|:Wr"- T6UpYОЊR D3]f0@?!a UЉW!!*H*1Z D9R{|sIw0X6}0{Cma=XAKs*ySTv X_CBq풩Yڭ>JF&yr={mʷ-umz9{ Z)[:ctIl&3vtziТ{1IQ5ZosF'a EtL9 v3X[N7j<VFSH!i#UMpO#~zvRi2-jtpZ h2@Kkk6 ȱAeS1`P;<3X!T= k %s'EUSl`)ԁ J>.DoN%E22ڬ(C5ArA %DF窈 8 Xu2ډaHFޮuTl}: Y yOP4GBBJGsMXP`^! E_rU'5zJHĘĒDUob@9#UAt YtZ ;cn@G&ɿA5fb b,fJΌ Pvә.0$b3PrBrTq2|9U@#/_҂h&5i̝5h|J1KWꨭO!Z.&riC"٣;(.+"S*!)4Y՘|(;lRH%;Іq(<>t@HU F+e>ڊ=N[1$NVA<*z4 jDBzY\'&I|^눭1yP*ASq`qkB.`N䢊Ɇ3mP!N{b9qħEf-k͋L 7ҍn2Ct(LQ"*CF"˪X"jc*z쎅#(.$ޡL撳b;Q;lbB@"Cq=PdɈ:dJ6aU.TS$[UlW :W~y$MGY(Nl+{F5J[ZLÁ -:i:i{A&[D郫ZN  jsqwz[9nzr|JY;qGddif;c=LΑ=%-^V_3a=}=iصmUMyUqz KݍR=,&F4MӞG3$'?򩎥kV1WCƎ4hiNƜ nR9v|3?|"%S-f%7٧8h8坖wL;mw XI˂^9X '-"ӷm" 붇Ql*ɫNSz}G'-sMSl?dnt2אo\D;d )g!iDmSȷa IiZPd[4 g0:>|`27FVD+v6p34kΔ&<\ 9>;2e"̛]DXRΆ0>{¦qPL\vu;@ka$]&) Mb}`,[daÚѪg*Vm"P7' שCp"rj7"u+PiZ0.0r\e K8PT9华 kb*T+R)kqOaAw@s=!+u`'KTdͩ_iq{C/+㡂pq{dɲ?[7;,үQDvvs]-FpoOm?L ϢiؙqƢ&b);{JS٨u~2{+- 4=饇j8v<[=^[n q7OoX7<]Zɭyt_RXǮш#3MMQaboI [my[m{[ex;&[ږ15M02L)=,7=,3e;q1l3]םc=xEd Iۋyg;}񩇈|hߋEi0K` jr5:9h92` X8M(Q"Uzᅹn/NhÙ.ЂC7&>i̟}J\ȓwZ#d+Yh )ژKH8_*ɫt&~ߧ6Et~( PhjAoZOD8yt`z$.ⅫrGVnUVkU ;|P~0T}! rg Orڮ\=|@OW7`/Q0[dnJ)<}4AK9~jdƱ`q[CBq)6+K46"yoErgZpȝ<3- -o8kDޝ9X7oH; qYpGÕ恔F<ªYt/g:ߥBWE;72&ᆇ\\LNUE\(\]tu}$kzyJ/kS~IGB!K[%j AdJCy`ʸк0?W>tdcL+P+6Z+䏮 "Hx`*˒n{V?Z  fipɑ"YiTMF^^lCh*KBO/cF?ûP_ _ŊIzC GX2ɑεADK|@Hb)_c]rю #q},^CI=z/2tqLƀnΆLv׮aIr)!9%0P0xMݤiBJcGpG,} ^wLad Zemgg7VGXW=3Be G&V3FW4RNHت|)>^p) L+GǍubXk7TD$p\b5p?B|'TJ`JJm1Y⨓%K&3jV`A,_̓zP-1U_Qt;\WM20`L/+ËMȩD{w]!&Gw5I;'rHAGcc*-YލQx%ՔWo ٳf4.z3|e5'tNh"ddv7 窈V MʝgdW+rb1tm?6{[8w^?)vcq*Y_'~y`YWWʹ&nv49y؀U7 /F8i'/x>: ?fSR&}?r#7T^"7RϬrFbtH @*A{,sp"-B$eN)E=qg*?ļ'ErHhՙeEMLSRAs-FE%iܤQ F2T";.~yN2vP]Yo#ɑ+^^<"g0bS:g$%)UؒZ bU_FFdFFpVw~(zoNr_iuȌцo?B1JJ9$Ϳ2a7GEDUwELgK|;7K#W׃?r@>$gFic7}ROߏ YkO ؚ륬{6O4ņikJB1Gnοk|;nadf|ͅ{@q56/V &ʲF=tᜆPY]~?jWLo \nd s ,+n4NiQq>ݜ?1$ƳR(-^`nW'+ _roe:W:~: djĻH~]\ `x(ϩ %#K*9 h* W( gizÝ٢ihYEØI8Xajħ>¾v, mX(%/cbߴ[2Hg@ٓ-ri[/Ȇ^$/ZVPt`,x|_g@=bk~w6,C d#;; >y*QZoX3ƶ`5d"FDUDKNٙhi Ōvo=a EnΛbokpŲsgjٹOu|xusnmnWW]jZ#ڌ.ׯۻ Ys\Ruݠ ޻*aٙp S#aSqD` ؎{#_D^1|G߉j%~FTr+wߞjOגjr5 z龳r"a;LtSښj {N5GswʠIwh%ҩuOn} C tGȡӏRhݺʠ t>u{>ߋt S޻nݓ-Z>! Sxcf8UMmsF5~u!k 1ŇBsktm[:mk;<Ѷd&M_m[oxJ Χ|vvh9N0[e}i;ݵ,Yؠc]%R%+*vS&=V{q6NvtVlҁ+uj8(]kM7bS Gr2\^}㣘Ϫx]ά|V ЁxJoc͏ݴjn&N Ye`0@8X{0ЯǴFd35+=Uq+z+ *izwɳW˔mrͅ7viHI)N`Z$FW=m L#imtՅ%&SMt ]oi MU9,> u?V5ֲktIewBfآ2ޗe6z[xNܿ|R&͞wX ,G:,w*cAFr(] oQ1X97AJAmxV-Xp (E,h1Ɛ ?ƂtFQEt(RBĆȆ"ANAf2[!kĜ':{YgIBƝօj"egs,S[l^eӢUTJP3qY(}iQ~?-e HzmXjG0Q](-VeJ֪.sn)2NL˺c5q ouՅj+dpouz ^zD{m4) ^AXn9ЁyLG!:/fZifOٔ11AP2! AFHQD e{6zFz=KO܀f3k??ȶXÅEgX38Vj )t uC%ȩPn;ͻuCE`#[@ocI~cwEh H2` h>HƌRZ0+K`egK[MJv B982eXV0vx#X:-/m[򰫏IH@hBy!h>qKo xzH7]vT5[ArNKB$dftd= U:)E)'̄ ?[ZfT % KS6Bfg%zDRA\J_'UFӖ,A(FgwڻcN{͔C4$cdbɍ6Z/FCFrT6SHsYȈ{TXf:EF=jD+Ȝtdni" YU@50[F,"MKR9dWF. 'l49#Rs¥qF7"sI$Д  kI@W * !"5p2ہM<J QI$H:(C4Y)qK/Kjbftj^K;D cA$ܭ ndZ|qZԿ>[d?v:<,\Oٽ/BD^yes?||TVk P Ŋ&,[qAFy/ ESvڝhjMm&W>tda3 2 +%6c<^DTQۧΜ+7;d7j5!3nhʪ0eWOY 4h5dչ̑W&L~}M(>NL7̣)!e,!vHy2>i5Va[|*M וX)Q"ώMU~U1@5Ike ,0-"tVBZ@H*gŁJB |qR<7ВK'^^te/nlQb|oΕaK~\,bKivl)PdXƮY#%>k9k/t|LPrc6*@V±{Y3T!q]7r Bӌ+YǛo֕wnQvZPmaR!x,X]~{s#E}@eWpxh:i [-%ѐ%A#&FDjіy;}IG0P qw Z8tڞj)-ݘf&٩d$Q9T.Y魀m`d!hB'άB5z ,#jokm*-lɎNq-ǸVh<-kTKՌj%Í%GBF$D#I (ȅ"Dm f(ziFNJt+9&,?.g660h|~g 4o"]\Ke@f&H*9L4]iuq2&GU%++ G8~Tv+QawnIZٯ\!a:H_urs~]i?חWO5~Bbt@LjW~[w`dgn^=]z|EX }-UM_\ߎ#e2(k:fbt}ƂIyNGdtñڰy랲Q,6 uNMy̵~u{yݟuCԺW5L醘CAOj}oces; ލ+蹽\On篋pPV(ZnC |"Dl-¼lL.M[JB6:3o+ȆuC@#L>&&Y|0ECU!R0(I&;i$f2m-ۚ\.FLJ238 !YMb:2mFmO52cں RV*#33JH%9"." βr R?{ƍ K/{vsµpi/I7[%qI*~!%(JpPeSLh_8{` &rHhBԽ5ZD]z^I & Q.4 ԭT% /Q3}jwr%yCďq yAvXNoWrg}Oߒ!s0zBrnr{_TյQoCr [wnAZFTv>'TK>:켲t0F +}et/TgZi5FpzNeJUrca(0˕Vd3JKj`^Jre%uqUys=6.FWo1"܅ Ne(0AAH#JbGO-yRtJ7T<[6^۠ظJ:VND-vVq,W*WCdwV]2@V!I/WуnFvÌ>PmHH޻Q4j2uէЃe/j=»5׃iz8խtBfzHoT vH4^:duY-~ J5 ut2Ӏ`5A;cG E#=<ΒKҭgIVwb,[gg:ǝTĄ (Вzz{ށ#֢85v~G }oqכ$NJrtUCy=~:aqmLI&2.꘢{m;d>%)EDI\c) EGL͓bE8ߕmӾ;]{;7#)C9dh4^ԋR' k&U2C%0 qk8oK8q[Q1TtTƅz0oqL<[=pĺߪn U 'T]?a'oAE?0rnK?]ۣN oIGRO?;Ne2-4u;Yǀo.~G!,fr:yy; ": yjH,hpחYUϋD \!2(I%(z+U9--Sы#[PQ yיG Zݦ ZeJZk0lTVlP|n' 5jҴC[<I/K4uvhi<h1>#ȧI;[9ђlyGyۈޘhwt.G1+I㽃c\Z烍7s O߱@ΛY7?:|Ni+q7l[j2:C ʾI].߷u-.Ոl?)Ptsթ uchZ8[<7)Dk!l=Fܘ@x:ȧ:Nk\}TWp9PS x 26 %Yڿ[>o4DPb[M1`]^ZmZ 6l}uB/9YZIw[ '}GpZe[ne{wa!߸u[lq{twޭ rL;xF3Q'Lfz.,76{q>nZrV9M.WZ*5־'ӻua!߸.̴ٔ)1KsYr1xC3)K ˊ$,?O2pHXeUt^(KHFh%VxpH4-V*:[*mKWI71YMtG?I4¡gܣ % >H95FM P%&PoW*BQ2/0Zh*a) ֔| |=c8F;.bP͍Ep ц\T39W聬4tTgN=,|Iޗ<;8x߽yJߔJ@T0=O#҇>̷RTD79s΃t&EЊsZ3g2XRю*{2]),"P6 8AkB2zDo8h 4YM^S8`J}$=-ϔ-M[P.hžB9mQh FP}15,,(rO[,64߃8|y/'q9;_Bk9Jg=\?s~5Np1j6!=7?Ƈwm]}_wv3F((9V&MBύJB 7& Ep 0)"+gՖ_[Neܵ`.E<|{+A+~o3TuUtV|M4D*beFHk*^;+k]sv4w&sK6S UaˡZ1"jru5kO\vXv7e kuگgF E d(t{F?U/ݥ}$ږlW4#_\io]+4;WIX6A#i1ؐF..6:Τ<蠑6ƳY)GJ1hr][|%jSڊRп qC-3জ/Ҿ2 L< ̹0~:!m"|k,OY&< vrvUX1D&Pg1OJSRh IÖZykAX `k0U\n7> b 6_8Qثg׶f%Sz۳G HI@#DRb’( dT+G8 S1xל_>Ua95x B2:#9(-!sPTJd1i>T"g5mB>RdJ'"S2n`,% sѓ2$BRM^)z|.*sHJ`*@9u5~~}UY*Tq{~TaZ=W~:Hg3xI(3o}Gya2-h祷G*-BT?i߹Zx"pm>,%l}{p; ¥NGHCCoQ`yJvd½8gd9Ց6~3 gBD} D +}e<`﹕̜ +9RHy=RxCft_o샕>O2 JJ)ɳRJR",[tdyd/R^3W=f[ϪDըKwek~1`@>QBHS]P-Ac2ĭ/G碩}-P@q.% 3 %. ƹyzD )9j]Gd?oKf4sWT#~{nWݶ@/<+Np+o+͌t)J.L1:b'hPNgEEp3JZ8 x9+/YsfF%I&&\MN5, Xh4N s,HrlEVIܖ/5ҵWBk-F]R.ɫSfJQ [!3ZY"s-!ȐHHE)PaMbiڌ Q67B - mF)W/\=6JY4`MYO&PPRc3eB" )%fGzCٹW.("-"\ )*Z lL[ǫ֖w D&Ty1Am;d-^Ha4;;eKB(+Q xaAElT>9J3 m0}cFf8? :VH#eT9-9Zo<}yL||y)tdajsTP@'K̛%Yc #`m,LxHT/ dj @bJVXڥ>GߩVVoHV boHΐ3JOY5p1ƣ$:Hв`08-vm?_>z1D:ByˋI10cSuI 8e=z^n:"8X h6[PQ`CI(#~ea}r™wwqô0_Ԩ6(m+4^`iTdǏķozŰ_fŀrVՒ)Mn6#3 {uqq}y6Z?݌oNq厡(㲢BBqo u\ TrԽ>&_޳m-ך~KڕTs{~U U\|[$p˙PH aڭ}i5J-\tѥ1 /]^b޵q#E^+8Pľp%H~7Ֆ]YJ?rWV&Rhcy goő@G^U<{Eh1z0h4F":3L11N)3h^ .[Qp[ʨ#d8]9ApBZ<B '!E,P1a5ٍoq$cFR1X1pbJB, d&> { (:rZT%ǒjU8`6UL!LڣtNܴͿĘ BKjjIK۹-=F-wt4ף%p>r7-ͩ%ȵ9nńJ4WTs^T#}}.7pT_RmhQk)7-ŐCbpҜj.e_zZ*T I\j&yK)jDP|i9PFz3cJ ap-H. 5KɮMeMwe'ki*Q焻MN |n>(Z{>CKm:Ѝ,o)6W?^ГZIvۯr-f潅v[E(#i](O΂OWͧvA3Q\VwW?p&@F0&VվkJwe<:Q~@}*{;cP?n٭g fp@ ܨ7yWDd:_1 W,z^N UuqGwWo&ft"c2z{8mwb/gxʃ_w:#((:w!8(ᚲFm ߦ\ڟ7ibiV]Zr{N}cSg%w$ʇBѐW 0x?o8[(\T'>֭ǽ ضu unMhȉhm"\ wŎV^j~t7۞V"ƯV.CNջw&O(PJ;dbgou63jg~{CsrJ)gcYr^A֮9I'[~uŁ+nK-׷63$f+~άV]S^!Rmd_oK/e[֩n 3} t~|,U>X)*b EW'Td˲{Fms̕9U^YuE8,|B(Ift1,EXLcƴbBHw5ڮ f1G bCg\ BH1*xf2V6.˸b3sP&б؟+Xo0xס :l\,ؒmMu`lQLG D6|@iE3aIH ʍ9t*GlF[_D|x?ٵ&ӣtYlkq|%g r !Roݠznd:E(W:(WMՙA< R٬ tQKYc-#>jA@t3n+`7\y),]>$PD  uS{D,yr\I@' B4GrܨGުAmX(k&Ϟ}(MGls:xϡ|"^^K wb 1_Jvd<8sx& NЁM;8ƇAD`|rqψWX3vU'^3R!l‹6+E;Î!;@K$Ȑ@^6 5iw9~s5sT+5󕖤CPԬxs)VD[UI *Oq6Jשg"}61Ԭ5(&"y-PN/*PPbZ= #(y,ʋwVf ;wLj6 کO5 x8ԋ'wxu'}M Q@(@ hȉhN)>֍ D`By:ntOm_ά[x*ukBCN\ECtJ=tܳns: }uaĸ3^hݚАW:EmFh-ƙO*|var/home/core/zuul-output/logs/kubelet.log0000644000000000000000003712735315140143756017716 0ustar rootrootFeb 02 15:15:15 crc systemd[1]: Starting Kubernetes Kubelet... Feb 02 15:15:15 crc restorecon[4700]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 02 15:15:15 crc restorecon[4700]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 02 15:15:16 crc restorecon[4700]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 02 15:15:16 crc restorecon[4700]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Feb 02 15:15:16 crc kubenswrapper[4733]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 02 15:15:16 crc kubenswrapper[4733]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Feb 02 15:15:16 crc kubenswrapper[4733]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 02 15:15:16 crc kubenswrapper[4733]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 02 15:15:16 crc kubenswrapper[4733]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Feb 02 15:15:16 crc kubenswrapper[4733]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 02 15:15:16 crc kubenswrapper[4733]: I0202 15:15:16.992085 4733 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002419 4733 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002460 4733 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002472 4733 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002483 4733 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002492 4733 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002501 4733 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002510 4733 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002518 4733 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002527 4733 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002535 4733 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002544 4733 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002553 4733 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002565 4733 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002576 4733 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002585 4733 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002594 4733 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002604 4733 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002613 4733 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002622 4733 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002631 4733 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002640 4733 feature_gate.go:330] unrecognized feature gate: Example Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002648 4733 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002657 4733 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002666 4733 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002674 4733 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002683 4733 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002691 4733 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002699 4733 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002707 4733 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002725 4733 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002734 4733 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002742 4733 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002751 4733 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002760 4733 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002769 4733 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002778 4733 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002787 4733 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002796 4733 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002805 4733 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002814 4733 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002824 4733 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002832 4733 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002842 4733 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002853 4733 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002862 4733 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002870 4733 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002879 4733 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002887 4733 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002895 4733 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002904 4733 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002912 4733 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002924 4733 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002934 4733 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002942 4733 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002951 4733 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002960 4733 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002968 4733 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002979 4733 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.002989 4733 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.003001 4733 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.003011 4733 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.003022 4733 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.003031 4733 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.003040 4733 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.003049 4733 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.003060 4733 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.003069 4733 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.003078 4733 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.003087 4733 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.003097 4733 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.003106 4733 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.006594 4733 flags.go:64] FLAG: --address="0.0.0.0" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.006642 4733 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.006660 4733 flags.go:64] FLAG: --anonymous-auth="true" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.006673 4733 flags.go:64] FLAG: --application-metrics-count-limit="100" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.006687 4733 flags.go:64] FLAG: --authentication-token-webhook="false" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.006699 4733 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.006712 4733 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.006724 4733 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.006735 4733 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.006744 4733 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.006755 4733 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.006770 4733 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.006781 4733 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.006791 4733 flags.go:64] FLAG: --cgroup-root="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.006800 4733 flags.go:64] FLAG: --cgroups-per-qos="true" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.006811 4733 flags.go:64] FLAG: --client-ca-file="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.006820 4733 flags.go:64] FLAG: --cloud-config="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.006830 4733 flags.go:64] FLAG: --cloud-provider="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.006839 4733 flags.go:64] FLAG: --cluster-dns="[]" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.006852 4733 flags.go:64] FLAG: --cluster-domain="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.006861 4733 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.006872 4733 flags.go:64] FLAG: --config-dir="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.006883 4733 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.006893 4733 flags.go:64] FLAG: --container-log-max-files="5" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.006907 4733 flags.go:64] FLAG: --container-log-max-size="10Mi" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.006918 4733 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.006929 4733 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.006939 4733 flags.go:64] FLAG: --containerd-namespace="k8s.io" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.006949 4733 flags.go:64] FLAG: --contention-profiling="false" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.006959 4733 flags.go:64] FLAG: --cpu-cfs-quota="true" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.006969 4733 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.006979 4733 flags.go:64] FLAG: --cpu-manager-policy="none" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.006989 4733 flags.go:64] FLAG: --cpu-manager-policy-options="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007001 4733 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007011 4733 flags.go:64] FLAG: --enable-controller-attach-detach="true" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007021 4733 flags.go:64] FLAG: --enable-debugging-handlers="true" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007031 4733 flags.go:64] FLAG: --enable-load-reader="false" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007041 4733 flags.go:64] FLAG: --enable-server="true" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007051 4733 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007063 4733 flags.go:64] FLAG: --event-burst="100" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007074 4733 flags.go:64] FLAG: --event-qps="50" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007084 4733 flags.go:64] FLAG: --event-storage-age-limit="default=0" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007094 4733 flags.go:64] FLAG: --event-storage-event-limit="default=0" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007104 4733 flags.go:64] FLAG: --eviction-hard="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007116 4733 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007126 4733 flags.go:64] FLAG: --eviction-minimum-reclaim="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007136 4733 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007147 4733 flags.go:64] FLAG: --eviction-soft="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007157 4733 flags.go:64] FLAG: --eviction-soft-grace-period="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007196 4733 flags.go:64] FLAG: --exit-on-lock-contention="false" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007206 4733 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007216 4733 flags.go:64] FLAG: --experimental-mounter-path="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007226 4733 flags.go:64] FLAG: --fail-cgroupv1="false" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007236 4733 flags.go:64] FLAG: --fail-swap-on="true" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007245 4733 flags.go:64] FLAG: --feature-gates="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007257 4733 flags.go:64] FLAG: --file-check-frequency="20s" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007267 4733 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007280 4733 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007290 4733 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007300 4733 flags.go:64] FLAG: --healthz-port="10248" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007310 4733 flags.go:64] FLAG: --help="false" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007320 4733 flags.go:64] FLAG: --hostname-override="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007330 4733 flags.go:64] FLAG: --housekeeping-interval="10s" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007340 4733 flags.go:64] FLAG: --http-check-frequency="20s" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007352 4733 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007362 4733 flags.go:64] FLAG: --image-credential-provider-config="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007371 4733 flags.go:64] FLAG: --image-gc-high-threshold="85" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007382 4733 flags.go:64] FLAG: --image-gc-low-threshold="80" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007393 4733 flags.go:64] FLAG: --image-service-endpoint="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007403 4733 flags.go:64] FLAG: --kernel-memcg-notification="false" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007413 4733 flags.go:64] FLAG: --kube-api-burst="100" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007423 4733 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007433 4733 flags.go:64] FLAG: --kube-api-qps="50" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007445 4733 flags.go:64] FLAG: --kube-reserved="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007455 4733 flags.go:64] FLAG: --kube-reserved-cgroup="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007465 4733 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007475 4733 flags.go:64] FLAG: --kubelet-cgroups="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007484 4733 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007494 4733 flags.go:64] FLAG: --lock-file="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007504 4733 flags.go:64] FLAG: --log-cadvisor-usage="false" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007514 4733 flags.go:64] FLAG: --log-flush-frequency="5s" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007524 4733 flags.go:64] FLAG: --log-json-info-buffer-size="0" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007552 4733 flags.go:64] FLAG: --log-json-split-stream="false" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007566 4733 flags.go:64] FLAG: --log-text-info-buffer-size="0" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007576 4733 flags.go:64] FLAG: --log-text-split-stream="false" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007587 4733 flags.go:64] FLAG: --logging-format="text" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007596 4733 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007608 4733 flags.go:64] FLAG: --make-iptables-util-chains="true" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007618 4733 flags.go:64] FLAG: --manifest-url="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007629 4733 flags.go:64] FLAG: --manifest-url-header="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007643 4733 flags.go:64] FLAG: --max-housekeeping-interval="15s" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007653 4733 flags.go:64] FLAG: --max-open-files="1000000" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007666 4733 flags.go:64] FLAG: --max-pods="110" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007676 4733 flags.go:64] FLAG: --maximum-dead-containers="-1" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007686 4733 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007696 4733 flags.go:64] FLAG: --memory-manager-policy="None" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007707 4733 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007717 4733 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007727 4733 flags.go:64] FLAG: --node-ip="192.168.126.11" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007739 4733 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007761 4733 flags.go:64] FLAG: --node-status-max-images="50" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007771 4733 flags.go:64] FLAG: --node-status-update-frequency="10s" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007781 4733 flags.go:64] FLAG: --oom-score-adj="-999" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007791 4733 flags.go:64] FLAG: --pod-cidr="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007801 4733 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007814 4733 flags.go:64] FLAG: --pod-manifest-path="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007824 4733 flags.go:64] FLAG: --pod-max-pids="-1" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007835 4733 flags.go:64] FLAG: --pods-per-core="0" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007844 4733 flags.go:64] FLAG: --port="10250" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007855 4733 flags.go:64] FLAG: --protect-kernel-defaults="false" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007864 4733 flags.go:64] FLAG: --provider-id="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007874 4733 flags.go:64] FLAG: --qos-reserved="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007884 4733 flags.go:64] FLAG: --read-only-port="10255" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007893 4733 flags.go:64] FLAG: --register-node="true" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007903 4733 flags.go:64] FLAG: --register-schedulable="true" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007913 4733 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007929 4733 flags.go:64] FLAG: --registry-burst="10" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007939 4733 flags.go:64] FLAG: --registry-qps="5" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007949 4733 flags.go:64] FLAG: --reserved-cpus="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007960 4733 flags.go:64] FLAG: --reserved-memory="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007979 4733 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.007989 4733 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.008000 4733 flags.go:64] FLAG: --rotate-certificates="false" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.008010 4733 flags.go:64] FLAG: --rotate-server-certificates="false" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.008019 4733 flags.go:64] FLAG: --runonce="false" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.008030 4733 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.008040 4733 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.008049 4733 flags.go:64] FLAG: --seccomp-default="false" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.008059 4733 flags.go:64] FLAG: --serialize-image-pulls="true" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.008069 4733 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.008080 4733 flags.go:64] FLAG: --storage-driver-db="cadvisor" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.008090 4733 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.008100 4733 flags.go:64] FLAG: --storage-driver-password="root" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.008110 4733 flags.go:64] FLAG: --storage-driver-secure="false" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.008120 4733 flags.go:64] FLAG: --storage-driver-table="stats" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.008130 4733 flags.go:64] FLAG: --storage-driver-user="root" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.008140 4733 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.008151 4733 flags.go:64] FLAG: --sync-frequency="1m0s" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.008186 4733 flags.go:64] FLAG: --system-cgroups="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.008196 4733 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.008212 4733 flags.go:64] FLAG: --system-reserved-cgroup="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.008222 4733 flags.go:64] FLAG: --tls-cert-file="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.008231 4733 flags.go:64] FLAG: --tls-cipher-suites="[]" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.008244 4733 flags.go:64] FLAG: --tls-min-version="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.008255 4733 flags.go:64] FLAG: --tls-private-key-file="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.008266 4733 flags.go:64] FLAG: --topology-manager-policy="none" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.008277 4733 flags.go:64] FLAG: --topology-manager-policy-options="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.008287 4733 flags.go:64] FLAG: --topology-manager-scope="container" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.008297 4733 flags.go:64] FLAG: --v="2" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.008309 4733 flags.go:64] FLAG: --version="false" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.008322 4733 flags.go:64] FLAG: --vmodule="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.008334 4733 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.008345 4733 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.008595 4733 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.008606 4733 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.008617 4733 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.008626 4733 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.008635 4733 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.008644 4733 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.008657 4733 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.008669 4733 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.008679 4733 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.008690 4733 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.008699 4733 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.008708 4733 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.008716 4733 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.008725 4733 feature_gate.go:330] unrecognized feature gate: Example Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.008734 4733 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.008743 4733 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.008752 4733 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.008761 4733 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.008769 4733 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.008778 4733 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.008787 4733 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.008796 4733 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.008805 4733 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.008813 4733 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.008822 4733 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.008830 4733 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.008839 4733 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.008849 4733 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.008858 4733 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.008866 4733 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.008875 4733 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.008889 4733 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.008897 4733 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.008906 4733 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.008914 4733 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.008925 4733 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.008936 4733 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.008944 4733 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.008954 4733 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.008963 4733 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.008971 4733 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.008980 4733 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.008988 4733 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.008996 4733 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.009005 4733 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.009014 4733 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.009022 4733 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.009031 4733 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.009039 4733 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.009050 4733 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.009061 4733 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.009071 4733 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.009083 4733 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.009094 4733 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.009105 4733 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.009113 4733 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.009122 4733 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.009131 4733 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.009139 4733 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.009148 4733 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.009156 4733 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.009190 4733 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.009199 4733 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.009214 4733 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.009222 4733 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.009231 4733 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.009240 4733 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.009249 4733 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.009257 4733 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.009265 4733 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.009274 4733 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.009299 4733 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.024126 4733 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.024215 4733 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024349 4733 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024363 4733 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024374 4733 feature_gate.go:330] unrecognized feature gate: Example Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024384 4733 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024393 4733 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024402 4733 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024411 4733 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024420 4733 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024430 4733 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024439 4733 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024448 4733 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024457 4733 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024466 4733 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024475 4733 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024484 4733 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024492 4733 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024501 4733 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024509 4733 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024519 4733 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024528 4733 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024536 4733 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024544 4733 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024553 4733 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024562 4733 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024570 4733 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024578 4733 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024587 4733 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024595 4733 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024604 4733 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024612 4733 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024621 4733 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024630 4733 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024638 4733 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024646 4733 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024658 4733 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024670 4733 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024683 4733 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024693 4733 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024701 4733 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024710 4733 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024719 4733 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024728 4733 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024737 4733 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024745 4733 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024755 4733 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024764 4733 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024772 4733 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024780 4733 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024789 4733 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024798 4733 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024807 4733 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024815 4733 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024823 4733 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024832 4733 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024840 4733 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024852 4733 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024863 4733 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024874 4733 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024885 4733 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024894 4733 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024904 4733 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024914 4733 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024922 4733 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024931 4733 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024939 4733 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024948 4733 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024957 4733 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024965 4733 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024976 4733 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024987 4733 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.024999 4733 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.025014 4733 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025323 4733 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025336 4733 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025346 4733 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025355 4733 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025365 4733 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025375 4733 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025384 4733 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025394 4733 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025405 4733 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025414 4733 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025423 4733 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025432 4733 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025440 4733 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025449 4733 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025458 4733 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025466 4733 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025475 4733 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025483 4733 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025491 4733 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025501 4733 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025509 4733 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025518 4733 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025526 4733 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025535 4733 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025544 4733 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025552 4733 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025561 4733 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025569 4733 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025578 4733 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025586 4733 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025595 4733 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025603 4733 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025612 4733 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025620 4733 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025642 4733 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025650 4733 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025661 4733 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025670 4733 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025678 4733 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025686 4733 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025695 4733 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025706 4733 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025717 4733 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025727 4733 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025736 4733 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025746 4733 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025755 4733 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025764 4733 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025772 4733 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025780 4733 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025789 4733 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025797 4733 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025806 4733 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025814 4733 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025822 4733 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025831 4733 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025839 4733 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025848 4733 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025856 4733 feature_gate.go:330] unrecognized feature gate: Example Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025864 4733 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025876 4733 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025887 4733 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025896 4733 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025906 4733 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025915 4733 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025926 4733 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025937 4733 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025947 4733 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025959 4733 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025971 4733 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.025994 4733 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.026009 4733 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.026352 4733 server.go:940] "Client rotation is on, will bootstrap in background" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.032619 4733 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.032745 4733 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.034474 4733 server.go:997] "Starting client certificate rotation" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.034524 4733 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.036261 4733 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-30 10:42:44.415508866 +0000 UTC Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.036385 4733 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.062212 4733 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.064918 4733 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 02 15:15:17 crc kubenswrapper[4733]: E0202 15:15:17.065947 4733 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.2:6443: connect: connection refused" logger="UnhandledError" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.089772 4733 log.go:25] "Validated CRI v1 runtime API" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.129471 4733 log.go:25] "Validated CRI v1 image API" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.133145 4733 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.138075 4733 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-02-02-15-11-04-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.138111 4733 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:45 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:42 fsType:tmpfs blockSize:0}] Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.157004 4733 manager.go:217] Machine: {Timestamp:2026-02-02 15:15:17.153941075 +0000 UTC m=+0.605402473 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:91b64b75-3cd9-4fb9-93f0-dbd3bcdcaec4 BootID:cd742269-802e-45ec-a88e-17aee17c574e Filesystems:[{Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:45 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:42 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:d1:14:f9 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:d1:14:f9 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:9d:26:2c Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:84:5b:69 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:9d:18:fa Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:64:27:4b Speed:-1 Mtu:1496} {Name:eth10 MacAddress:ea:bb:a1:c7:40:06 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:1e:4f:2f:77:de:b9 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.157352 4733 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.157514 4733 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.160180 4733 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.160505 4733 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.160566 4733 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.160891 4733 topology_manager.go:138] "Creating topology manager with none policy" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.160910 4733 container_manager_linux.go:303] "Creating device plugin manager" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.161411 4733 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.161465 4733 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.161713 4733 state_mem.go:36] "Initialized new in-memory state store" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.161843 4733 server.go:1245] "Using root directory" path="/var/lib/kubelet" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.166259 4733 kubelet.go:418] "Attempting to sync node with API server" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.166293 4733 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.166357 4733 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.166379 4733 kubelet.go:324] "Adding apiserver pod source" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.166403 4733 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.171086 4733 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.2:6443: connect: connection refused Feb 02 15:15:17 crc kubenswrapper[4733]: E0202 15:15:17.171240 4733 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.2:6443: connect: connection refused" logger="UnhandledError" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.171266 4733 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.171314 4733 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.2:6443: connect: connection refused Feb 02 15:15:17 crc kubenswrapper[4733]: E0202 15:15:17.171413 4733 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.2:6443: connect: connection refused" logger="UnhandledError" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.172513 4733 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.175548 4733 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.177288 4733 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.177329 4733 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.177344 4733 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.177357 4733 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.177379 4733 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.177393 4733 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.177407 4733 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.177429 4733 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.177446 4733 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.177463 4733 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.177481 4733 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.177495 4733 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.179240 4733 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.179819 4733 server.go:1280] "Started kubelet" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.180867 4733 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Feb 02 15:15:17 crc systemd[1]: Started Kubernetes Kubelet. Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.181716 4733 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.180883 4733 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.187289 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.2:6443: connect: connection refused Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.188294 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.188407 4733 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.189270 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 14:18:09.512360181 +0000 UTC Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.189841 4733 volume_manager.go:287] "The desired_state_of_world populator starts" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.189880 4733 volume_manager.go:289] "Starting Kubelet Volume Manager" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.190043 4733 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Feb 02 15:15:17 crc kubenswrapper[4733]: E0202 15:15:17.191049 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.192290 4733 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.192326 4733 factory.go:55] Registering systemd factory Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.192341 4733 factory.go:221] Registration of the systemd container factory successfully Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.196076 4733 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.2:6443: connect: connection refused Feb 02 15:15:17 crc kubenswrapper[4733]: E0202 15:15:17.196154 4733 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.2:6443: connect: connection refused" logger="UnhandledError" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.196370 4733 server.go:460] "Adding debug handlers to kubelet server" Feb 02 15:15:17 crc kubenswrapper[4733]: E0202 15:15:17.196782 4733 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.2:6443: connect: connection refused" interval="200ms" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.197144 4733 factory.go:153] Registering CRI-O factory Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.197207 4733 factory.go:221] Registration of the crio container factory successfully Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.197254 4733 factory.go:103] Registering Raw factory Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.197277 4733 manager.go:1196] Started watching for new ooms in manager Feb 02 15:15:17 crc kubenswrapper[4733]: E0202 15:15:17.196318 4733 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.2:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.189076d67ed80fd3 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-02 15:15:17.179772883 +0000 UTC m=+0.631234281,LastTimestamp:2026-02-02 15:15:17.179772883 +0000 UTC m=+0.631234281,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.198214 4733 manager.go:319] Starting recovery of all containers Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.210983 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.211109 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.211133 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.211154 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.211257 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.211276 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.211296 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.211313 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.211333 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.211352 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.211379 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.211397 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.211415 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.211435 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.211452 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.211470 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.211492 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.211510 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.211528 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.211548 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.211566 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.211586 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.211607 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.211625 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.211642 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.211660 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.211684 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.211704 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.211722 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.211739 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.211767 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.211808 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.211851 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.211874 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.211899 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.211921 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.211943 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.211967 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.211990 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.212012 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.212036 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.212062 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.212084 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.212107 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.212131 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.212158 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.212224 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.212251 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.212276 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.212301 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.212324 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.212351 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.212393 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.212421 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.212448 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.212474 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.212502 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.212527 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.212550 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.212573 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.212595 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.212616 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.212639 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.212660 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.212686 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.212710 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.212734 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.212757 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.212784 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.212807 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.212830 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.212854 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.212877 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.212919 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.213030 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.213052 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.213071 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.213105 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.213124 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.213142 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.213183 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.213201 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.213221 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.213237 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.213253 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.213271 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.213292 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.213309 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.213328 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.213344 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.213361 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.213379 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.213397 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.213416 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.213435 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.213453 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.213471 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.213491 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.213509 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.213527 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.213545 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.213564 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.213582 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.213600 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.213627 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.213648 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.213667 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.213688 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.213707 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.213727 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.213750 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.213768 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.213792 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.213810 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.213828 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.213846 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.213864 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.213882 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.213899 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.213918 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.213938 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.213956 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.213974 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.213992 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.214009 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.214026 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.214044 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.214061 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.214080 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.214097 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.214115 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.214132 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.214151 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.214239 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.214256 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.214272 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.214289 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.214305 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.214323 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.214340 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.214358 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.214374 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.214393 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.214411 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.214429 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.214446 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.214464 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.214481 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.214498 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.214516 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.214535 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.214604 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.214644 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.214687 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.214711 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.214775 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.214802 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.214826 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.214853 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.214878 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.214904 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.214927 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.214948 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.214973 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.214995 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.215018 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.215041 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.222338 4733 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.223043 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.223128 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.223210 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.223244 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.223272 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.223312 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.223340 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.223376 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.223406 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.223435 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.223472 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.223503 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.223539 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.223566 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.223596 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.223632 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.223661 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.223697 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.223730 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.223759 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.223795 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.223823 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.223859 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.223885 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.223910 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.223944 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.223975 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.224026 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.224061 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.224093 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.224134 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.224211 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.224255 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.224287 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.224319 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.224353 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.224384 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.224425 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.224457 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.224488 4733 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.224516 4733 reconstruct.go:97] "Volume reconstruction finished" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.224542 4733 reconciler.go:26] "Reconciler: start to sync state" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.224766 4733 manager.go:324] Recovery completed Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.242494 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.244961 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.245127 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.245158 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.246851 4733 cpu_manager.go:225] "Starting CPU manager" policy="none" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.246884 4733 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.246915 4733 state_mem.go:36] "Initialized new in-memory state store" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.251460 4733 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.253707 4733 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.253777 4733 status_manager.go:217] "Starting to sync pod status with apiserver" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.253814 4733 kubelet.go:2335] "Starting kubelet main sync loop" Feb 02 15:15:17 crc kubenswrapper[4733]: E0202 15:15:17.253890 4733 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.255500 4733 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.2:6443: connect: connection refused Feb 02 15:15:17 crc kubenswrapper[4733]: E0202 15:15:17.255594 4733 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.2:6443: connect: connection refused" logger="UnhandledError" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.271688 4733 policy_none.go:49] "None policy: Start" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.273042 4733 memory_manager.go:170] "Starting memorymanager" policy="None" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.273074 4733 state_mem.go:35] "Initializing new in-memory state store" Feb 02 15:15:17 crc kubenswrapper[4733]: E0202 15:15:17.291572 4733 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.322089 4733 manager.go:334] "Starting Device Plugin manager" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.322299 4733 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.322386 4733 server.go:79] "Starting device plugin registration server" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.322929 4733 eviction_manager.go:189] "Eviction manager: starting control loop" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.323061 4733 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.323340 4733 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.323498 4733 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.323576 4733 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Feb 02 15:15:17 crc kubenswrapper[4733]: E0202 15:15:17.335451 4733 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.354890 4733 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc"] Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.355019 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.357679 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.357729 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.357747 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.357969 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.358260 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.358338 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.359282 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.359343 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.359368 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.359538 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.359666 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.359704 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.359723 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.359729 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.359784 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.360810 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.360843 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.360858 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.361017 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.361217 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.361276 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.361825 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.361863 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.361882 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.361940 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.361962 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.361976 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.362113 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.362235 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.362263 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.363040 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.363075 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.363089 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.363209 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.363253 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.363277 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.363895 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.363930 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.363948 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.364212 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.364252 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.365129 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.365232 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.365251 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:17 crc kubenswrapper[4733]: E0202 15:15:17.398526 4733 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.2:6443: connect: connection refused" interval="400ms" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.424118 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.425293 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.425348 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.425370 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.425403 4733 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 02 15:15:17 crc kubenswrapper[4733]: E0202 15:15:17.425940 4733 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.2:6443: connect: connection refused" node="crc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.426592 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.426649 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.426685 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.426718 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.426779 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.426833 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.426880 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.426903 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.426968 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.426989 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.427066 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.427211 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.427249 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.427279 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.427360 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.528710 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.528770 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.528957 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.528992 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.528793 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.529086 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.529131 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.529138 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.529229 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.529187 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.529277 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.529316 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.529292 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.529309 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.529358 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.529424 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.529398 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.529513 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.529446 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.529560 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.529580 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.529599 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.529619 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.529654 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.529685 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.529638 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.529715 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.529741 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.529721 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.529846 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.626479 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.628249 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.628330 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.628354 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.628399 4733 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 02 15:15:17 crc kubenswrapper[4733]: E0202 15:15:17.629084 4733 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.2:6443: connect: connection refused" node="crc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.717715 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.725441 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.749923 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.775825 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-f23a0583111aa316ae3e252b3a645bd07d8e84022dc03149b04aeb97ab59a632 WatchSource:0}: Error finding container f23a0583111aa316ae3e252b3a645bd07d8e84022dc03149b04aeb97ab59a632: Status 404 returned error can't find the container with id f23a0583111aa316ae3e252b3a645bd07d8e84022dc03149b04aeb97ab59a632 Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.778137 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-e2a032592a1593bc20c0fc3dfc00ac55d021b134d429754021ff272f129a95e9 WatchSource:0}: Error finding container e2a032592a1593bc20c0fc3dfc00ac55d021b134d429754021ff272f129a95e9: Status 404 returned error can't find the container with id e2a032592a1593bc20c0fc3dfc00ac55d021b134d429754021ff272f129a95e9 Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.779386 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.784559 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-807bd1b0edb021f812bc4d419facae5469f029d5dd004ec26767d20e5b5bd220 WatchSource:0}: Error finding container 807bd1b0edb021f812bc4d419facae5469f029d5dd004ec26767d20e5b5bd220: Status 404 returned error can't find the container with id 807bd1b0edb021f812bc4d419facae5469f029d5dd004ec26767d20e5b5bd220 Feb 02 15:15:17 crc kubenswrapper[4733]: I0202 15:15:17.787252 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Feb 02 15:15:17 crc kubenswrapper[4733]: E0202 15:15:17.799846 4733 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.2:6443: connect: connection refused" interval="800ms" Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.803096 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-0221fd7dd1ad5ed9a68a72d137630f616571854fe62cb69558b93b8df80b7fca WatchSource:0}: Error finding container 0221fd7dd1ad5ed9a68a72d137630f616571854fe62cb69558b93b8df80b7fca: Status 404 returned error can't find the container with id 0221fd7dd1ad5ed9a68a72d137630f616571854fe62cb69558b93b8df80b7fca Feb 02 15:15:17 crc kubenswrapper[4733]: W0202 15:15:17.804078 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-ca0edfc4d6af351bf521aca62afd29689afd16546af0c70e95b2b525672e1775 WatchSource:0}: Error finding container ca0edfc4d6af351bf521aca62afd29689afd16546af0c70e95b2b525672e1775: Status 404 returned error can't find the container with id ca0edfc4d6af351bf521aca62afd29689afd16546af0c70e95b2b525672e1775 Feb 02 15:15:18 crc kubenswrapper[4733]: I0202 15:15:18.029397 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 15:15:18 crc kubenswrapper[4733]: I0202 15:15:18.031025 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:18 crc kubenswrapper[4733]: I0202 15:15:18.031066 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:18 crc kubenswrapper[4733]: I0202 15:15:18.031080 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:18 crc kubenswrapper[4733]: I0202 15:15:18.031113 4733 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 02 15:15:18 crc kubenswrapper[4733]: E0202 15:15:18.031600 4733 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.2:6443: connect: connection refused" node="crc" Feb 02 15:15:18 crc kubenswrapper[4733]: W0202 15:15:18.176435 4733 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.2:6443: connect: connection refused Feb 02 15:15:18 crc kubenswrapper[4733]: E0202 15:15:18.176533 4733 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.2:6443: connect: connection refused" logger="UnhandledError" Feb 02 15:15:18 crc kubenswrapper[4733]: I0202 15:15:18.189445 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 05:38:50.770178851 +0000 UTC Feb 02 15:15:18 crc kubenswrapper[4733]: I0202 15:15:18.189499 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.2:6443: connect: connection refused Feb 02 15:15:18 crc kubenswrapper[4733]: I0202 15:15:18.258391 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"807bd1b0edb021f812bc4d419facae5469f029d5dd004ec26767d20e5b5bd220"} Feb 02 15:15:18 crc kubenswrapper[4733]: I0202 15:15:18.259600 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e2a032592a1593bc20c0fc3dfc00ac55d021b134d429754021ff272f129a95e9"} Feb 02 15:15:18 crc kubenswrapper[4733]: I0202 15:15:18.261060 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f23a0583111aa316ae3e252b3a645bd07d8e84022dc03149b04aeb97ab59a632"} Feb 02 15:15:18 crc kubenswrapper[4733]: I0202 15:15:18.262363 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ca0edfc4d6af351bf521aca62afd29689afd16546af0c70e95b2b525672e1775"} Feb 02 15:15:18 crc kubenswrapper[4733]: I0202 15:15:18.263439 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"0221fd7dd1ad5ed9a68a72d137630f616571854fe62cb69558b93b8df80b7fca"} Feb 02 15:15:18 crc kubenswrapper[4733]: W0202 15:15:18.265306 4733 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.2:6443: connect: connection refused Feb 02 15:15:18 crc kubenswrapper[4733]: E0202 15:15:18.265403 4733 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.2:6443: connect: connection refused" logger="UnhandledError" Feb 02 15:15:18 crc kubenswrapper[4733]: W0202 15:15:18.516135 4733 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.2:6443: connect: connection refused Feb 02 15:15:18 crc kubenswrapper[4733]: E0202 15:15:18.516264 4733 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.2:6443: connect: connection refused" logger="UnhandledError" Feb 02 15:15:18 crc kubenswrapper[4733]: E0202 15:15:18.601070 4733 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.2:6443: connect: connection refused" interval="1.6s" Feb 02 15:15:18 crc kubenswrapper[4733]: W0202 15:15:18.621820 4733 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.2:6443: connect: connection refused Feb 02 15:15:18 crc kubenswrapper[4733]: E0202 15:15:18.621943 4733 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.2:6443: connect: connection refused" logger="UnhandledError" Feb 02 15:15:18 crc kubenswrapper[4733]: I0202 15:15:18.832661 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 15:15:18 crc kubenswrapper[4733]: I0202 15:15:18.833873 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:18 crc kubenswrapper[4733]: I0202 15:15:18.833915 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:18 crc kubenswrapper[4733]: I0202 15:15:18.833927 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:18 crc kubenswrapper[4733]: I0202 15:15:18.833952 4733 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 02 15:15:18 crc kubenswrapper[4733]: E0202 15:15:18.834447 4733 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.2:6443: connect: connection refused" node="crc" Feb 02 15:15:19 crc kubenswrapper[4733]: I0202 15:15:19.189567 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-05 10:20:41.06520839 +0000 UTC Feb 02 15:15:19 crc kubenswrapper[4733]: I0202 15:15:19.189941 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.2:6443: connect: connection refused Feb 02 15:15:19 crc kubenswrapper[4733]: I0202 15:15:19.221430 4733 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 02 15:15:19 crc kubenswrapper[4733]: E0202 15:15:19.222728 4733 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.2:6443: connect: connection refused" logger="UnhandledError" Feb 02 15:15:19 crc kubenswrapper[4733]: I0202 15:15:19.268661 4733 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71" exitCode=0 Feb 02 15:15:19 crc kubenswrapper[4733]: I0202 15:15:19.269253 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 15:15:19 crc kubenswrapper[4733]: I0202 15:15:19.269240 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71"} Feb 02 15:15:19 crc kubenswrapper[4733]: I0202 15:15:19.270650 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:19 crc kubenswrapper[4733]: I0202 15:15:19.270694 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:19 crc kubenswrapper[4733]: I0202 15:15:19.270725 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:19 crc kubenswrapper[4733]: I0202 15:15:19.273369 4733 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="5e8b7ecb2cfad9071bb3cc7968db64e99dc2fc24df7833a21ef302e062b5d3ce" exitCode=0 Feb 02 15:15:19 crc kubenswrapper[4733]: I0202 15:15:19.273449 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"5e8b7ecb2cfad9071bb3cc7968db64e99dc2fc24df7833a21ef302e062b5d3ce"} Feb 02 15:15:19 crc kubenswrapper[4733]: I0202 15:15:19.273507 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 15:15:19 crc kubenswrapper[4733]: I0202 15:15:19.274984 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:19 crc kubenswrapper[4733]: I0202 15:15:19.275032 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:19 crc kubenswrapper[4733]: I0202 15:15:19.275049 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:19 crc kubenswrapper[4733]: I0202 15:15:19.275262 4733 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="742ee25d4d54068f173176593cafc678ec8502b53656fe9c969177f803836e29" exitCode=0 Feb 02 15:15:19 crc kubenswrapper[4733]: I0202 15:15:19.275346 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"742ee25d4d54068f173176593cafc678ec8502b53656fe9c969177f803836e29"} Feb 02 15:15:19 crc kubenswrapper[4733]: I0202 15:15:19.275405 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 15:15:19 crc kubenswrapper[4733]: I0202 15:15:19.276667 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:19 crc kubenswrapper[4733]: I0202 15:15:19.276686 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:19 crc kubenswrapper[4733]: I0202 15:15:19.276694 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:19 crc kubenswrapper[4733]: I0202 15:15:19.278468 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"11600ada13fbc7db1831de385202bc1f041efb656cc58aa49b0f71fe902db9bb"} Feb 02 15:15:19 crc kubenswrapper[4733]: I0202 15:15:19.278517 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"2d5b032497b592363a03f89cb31ba1b551f479004072dddffce540c41190ab81"} Feb 02 15:15:19 crc kubenswrapper[4733]: I0202 15:15:19.278539 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"7a3fd02d55a10647428a05a9f53a1429a9827059c523d5deab3c4581c8f1b933"} Feb 02 15:15:19 crc kubenswrapper[4733]: I0202 15:15:19.279925 4733 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18" exitCode=0 Feb 02 15:15:19 crc kubenswrapper[4733]: I0202 15:15:19.279954 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18"} Feb 02 15:15:19 crc kubenswrapper[4733]: I0202 15:15:19.280077 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 15:15:19 crc kubenswrapper[4733]: I0202 15:15:19.280975 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:19 crc kubenswrapper[4733]: I0202 15:15:19.280999 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:19 crc kubenswrapper[4733]: I0202 15:15:19.281012 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:19 crc kubenswrapper[4733]: I0202 15:15:19.284962 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 15:15:19 crc kubenswrapper[4733]: I0202 15:15:19.286423 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:19 crc kubenswrapper[4733]: I0202 15:15:19.286463 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:19 crc kubenswrapper[4733]: I0202 15:15:19.286476 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:20 crc kubenswrapper[4733]: I0202 15:15:20.189325 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.2:6443: connect: connection refused Feb 02 15:15:20 crc kubenswrapper[4733]: I0202 15:15:20.190994 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-27 04:08:26.874492367 +0000 UTC Feb 02 15:15:20 crc kubenswrapper[4733]: E0202 15:15:20.202279 4733 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.2:6443: connect: connection refused" interval="3.2s" Feb 02 15:15:20 crc kubenswrapper[4733]: I0202 15:15:20.287975 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"226f5949820d4f498dbf6fa251c6edf1c6356900bd946e6afae2c29042b1c835"} Feb 02 15:15:20 crc kubenswrapper[4733]: I0202 15:15:20.288030 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f037e379a109ce230098e3ca223fd1dea3149de3ebe601f56e2bb295f3093a3c"} Feb 02 15:15:20 crc kubenswrapper[4733]: I0202 15:15:20.288047 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f6fdd19dac506f2d5d7eedc44ec2cda92a9381b73e91a45d949e4fcde4dfc283"} Feb 02 15:15:20 crc kubenswrapper[4733]: I0202 15:15:20.288060 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"fd383d69b9b27d52fa8672ff6e0294a74876d676973bdc0ca95125fef5f2d085"} Feb 02 15:15:20 crc kubenswrapper[4733]: I0202 15:15:20.290284 4733 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf" exitCode=0 Feb 02 15:15:20 crc kubenswrapper[4733]: I0202 15:15:20.290351 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf"} Feb 02 15:15:20 crc kubenswrapper[4733]: I0202 15:15:20.292365 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 15:15:20 crc kubenswrapper[4733]: I0202 15:15:20.293472 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 15:15:20 crc kubenswrapper[4733]: I0202 15:15:20.294704 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:20 crc kubenswrapper[4733]: I0202 15:15:20.294749 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:20 crc kubenswrapper[4733]: I0202 15:15:20.294762 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:20 crc kubenswrapper[4733]: I0202 15:15:20.295457 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"c25f4250ab2773e407dd24b8b81e0f4bf800d2cf3f7d5c3f50ffbd7f4c755c95"} Feb 02 15:15:20 crc kubenswrapper[4733]: I0202 15:15:20.296099 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:20 crc kubenswrapper[4733]: I0202 15:15:20.296128 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:20 crc kubenswrapper[4733]: I0202 15:15:20.296145 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:20 crc kubenswrapper[4733]: I0202 15:15:20.299332 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"cdff2efc8ee6095385be74fbe1bf30926e526639dcc00b39fca8d5c9a0145d3f"} Feb 02 15:15:20 crc kubenswrapper[4733]: I0202 15:15:20.299431 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"73fee4346d0af07556e8a1ea336818acea1f39aa8f703056b5954c7fac72aee1"} Feb 02 15:15:20 crc kubenswrapper[4733]: I0202 15:15:20.299442 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 15:15:20 crc kubenswrapper[4733]: I0202 15:15:20.299520 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"2ae3c8e412002f52fe150522d9bacb6415308ea41825b90d9d89010a7eca8764"} Feb 02 15:15:20 crc kubenswrapper[4733]: I0202 15:15:20.301991 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:20 crc kubenswrapper[4733]: I0202 15:15:20.302018 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:20 crc kubenswrapper[4733]: I0202 15:15:20.302032 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:20 crc kubenswrapper[4733]: I0202 15:15:20.304294 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"8fd39f5a05dec7a756dfd519f5df3ef7c77bbb5b8633c5af18a09ff8029cb079"} Feb 02 15:15:20 crc kubenswrapper[4733]: I0202 15:15:20.304633 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 15:15:20 crc kubenswrapper[4733]: I0202 15:15:20.306001 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:20 crc kubenswrapper[4733]: I0202 15:15:20.306046 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:20 crc kubenswrapper[4733]: I0202 15:15:20.306063 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:20 crc kubenswrapper[4733]: W0202 15:15:20.337039 4733 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.2:6443: connect: connection refused Feb 02 15:15:20 crc kubenswrapper[4733]: E0202 15:15:20.337140 4733 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.2:6443: connect: connection refused" logger="UnhandledError" Feb 02 15:15:20 crc kubenswrapper[4733]: W0202 15:15:20.419508 4733 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.2:6443: connect: connection refused Feb 02 15:15:20 crc kubenswrapper[4733]: E0202 15:15:20.419668 4733 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.2:6443: connect: connection refused" logger="UnhandledError" Feb 02 15:15:20 crc kubenswrapper[4733]: I0202 15:15:20.435088 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 15:15:20 crc kubenswrapper[4733]: I0202 15:15:20.436336 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:20 crc kubenswrapper[4733]: I0202 15:15:20.436376 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:20 crc kubenswrapper[4733]: I0202 15:15:20.436391 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:20 crc kubenswrapper[4733]: I0202 15:15:20.436422 4733 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 02 15:15:20 crc kubenswrapper[4733]: E0202 15:15:20.436869 4733 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.2:6443: connect: connection refused" node="crc" Feb 02 15:15:21 crc kubenswrapper[4733]: I0202 15:15:21.191456 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 02:36:32.45950109 +0000 UTC Feb 02 15:15:21 crc kubenswrapper[4733]: I0202 15:15:21.312145 4733 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5" exitCode=0 Feb 02 15:15:21 crc kubenswrapper[4733]: I0202 15:15:21.312279 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5"} Feb 02 15:15:21 crc kubenswrapper[4733]: I0202 15:15:21.312377 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 15:15:21 crc kubenswrapper[4733]: I0202 15:15:21.313919 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:21 crc kubenswrapper[4733]: I0202 15:15:21.313987 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:21 crc kubenswrapper[4733]: I0202 15:15:21.314013 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:21 crc kubenswrapper[4733]: I0202 15:15:21.317696 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 15:15:21 crc kubenswrapper[4733]: I0202 15:15:21.317697 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"992f4198192cbe4b1c89d575aa4a20c29c1115fd5eaa8986c6ed70c1d6b143c4"} Feb 02 15:15:21 crc kubenswrapper[4733]: I0202 15:15:21.317732 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 15:15:21 crc kubenswrapper[4733]: I0202 15:15:21.317727 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 15:15:21 crc kubenswrapper[4733]: I0202 15:15:21.318941 4733 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 02 15:15:21 crc kubenswrapper[4733]: I0202 15:15:21.319019 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 15:15:21 crc kubenswrapper[4733]: I0202 15:15:21.319731 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:21 crc kubenswrapper[4733]: I0202 15:15:21.319817 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:21 crc kubenswrapper[4733]: I0202 15:15:21.319842 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:21 crc kubenswrapper[4733]: I0202 15:15:21.320308 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:21 crc kubenswrapper[4733]: I0202 15:15:21.320360 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:21 crc kubenswrapper[4733]: I0202 15:15:21.320385 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:21 crc kubenswrapper[4733]: I0202 15:15:21.320578 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:21 crc kubenswrapper[4733]: I0202 15:15:21.320620 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:21 crc kubenswrapper[4733]: I0202 15:15:21.320639 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:21 crc kubenswrapper[4733]: I0202 15:15:21.320676 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:21 crc kubenswrapper[4733]: I0202 15:15:21.320707 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:21 crc kubenswrapper[4733]: I0202 15:15:21.320725 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:21 crc kubenswrapper[4733]: I0202 15:15:21.981962 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 15:15:21 crc kubenswrapper[4733]: I0202 15:15:21.991561 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 15:15:22 crc kubenswrapper[4733]: I0202 15:15:22.191794 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-04 00:51:46.912301881 +0000 UTC Feb 02 15:15:22 crc kubenswrapper[4733]: I0202 15:15:22.329553 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c08b087d85be9bb83007c7e38a5de9182a5091716fb8a2ca5ded1b8191bebc40"} Feb 02 15:15:22 crc kubenswrapper[4733]: I0202 15:15:22.329618 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"6c11b1859bf3bc0ec4d5e6a58d9becc1ac6333a1ed9820d83d634f1e93221c86"} Feb 02 15:15:22 crc kubenswrapper[4733]: I0202 15:15:22.329642 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"41d4a7bdc21f00b5039385d7227d8cb0f4ea3051d3b5065c71a49cb0bfad26a4"} Feb 02 15:15:22 crc kubenswrapper[4733]: I0202 15:15:22.329661 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"59175c8e5a7f3536168a69fd944bf3d536bcd7281eee96002343b7836a54401b"} Feb 02 15:15:22 crc kubenswrapper[4733]: I0202 15:15:22.329691 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 15:15:22 crc kubenswrapper[4733]: I0202 15:15:22.329644 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 15:15:22 crc kubenswrapper[4733]: I0202 15:15:22.329644 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 15:15:22 crc kubenswrapper[4733]: I0202 15:15:22.330764 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:22 crc kubenswrapper[4733]: I0202 15:15:22.330787 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:22 crc kubenswrapper[4733]: I0202 15:15:22.330797 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:22 crc kubenswrapper[4733]: I0202 15:15:22.332049 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:22 crc kubenswrapper[4733]: I0202 15:15:22.332199 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:22 crc kubenswrapper[4733]: I0202 15:15:22.332318 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:23 crc kubenswrapper[4733]: I0202 15:15:23.192718 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-12 19:01:43.936771214 +0000 UTC Feb 02 15:15:23 crc kubenswrapper[4733]: I0202 15:15:23.267701 4733 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 02 15:15:23 crc kubenswrapper[4733]: I0202 15:15:23.338442 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"5fd4969f0f16c9aa44208d2bf578f77d9d50914a8cb1764507175aa2eb58e733"} Feb 02 15:15:23 crc kubenswrapper[4733]: I0202 15:15:23.338488 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 15:15:23 crc kubenswrapper[4733]: I0202 15:15:23.338540 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 15:15:23 crc kubenswrapper[4733]: I0202 15:15:23.340154 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:23 crc kubenswrapper[4733]: I0202 15:15:23.340232 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:23 crc kubenswrapper[4733]: I0202 15:15:23.340252 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:23 crc kubenswrapper[4733]: I0202 15:15:23.340257 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:23 crc kubenswrapper[4733]: I0202 15:15:23.340286 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:23 crc kubenswrapper[4733]: I0202 15:15:23.340303 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:23 crc kubenswrapper[4733]: I0202 15:15:23.637602 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 15:15:23 crc kubenswrapper[4733]: I0202 15:15:23.639433 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:23 crc kubenswrapper[4733]: I0202 15:15:23.639532 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:23 crc kubenswrapper[4733]: I0202 15:15:23.639557 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:23 crc kubenswrapper[4733]: I0202 15:15:23.639606 4733 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 02 15:15:23 crc kubenswrapper[4733]: I0202 15:15:23.947715 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 02 15:15:23 crc kubenswrapper[4733]: I0202 15:15:23.948019 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 15:15:23 crc kubenswrapper[4733]: I0202 15:15:23.949508 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:23 crc kubenswrapper[4733]: I0202 15:15:23.949628 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:23 crc kubenswrapper[4733]: I0202 15:15:23.949649 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:23 crc kubenswrapper[4733]: I0202 15:15:23.959194 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 15:15:23 crc kubenswrapper[4733]: I0202 15:15:23.959331 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 15:15:23 crc kubenswrapper[4733]: I0202 15:15:23.960526 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:23 crc kubenswrapper[4733]: I0202 15:15:23.960571 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:23 crc kubenswrapper[4733]: I0202 15:15:23.960588 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:24 crc kubenswrapper[4733]: I0202 15:15:24.095272 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Feb 02 15:15:24 crc kubenswrapper[4733]: I0202 15:15:24.193482 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-12 00:50:10.709514338 +0000 UTC Feb 02 15:15:24 crc kubenswrapper[4733]: I0202 15:15:24.342843 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 15:15:24 crc kubenswrapper[4733]: I0202 15:15:24.345293 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:24 crc kubenswrapper[4733]: I0202 15:15:24.345357 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:24 crc kubenswrapper[4733]: I0202 15:15:24.345375 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:24 crc kubenswrapper[4733]: I0202 15:15:24.496826 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 15:15:24 crc kubenswrapper[4733]: I0202 15:15:24.497100 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 15:15:24 crc kubenswrapper[4733]: I0202 15:15:24.498827 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:24 crc kubenswrapper[4733]: I0202 15:15:24.498875 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:24 crc kubenswrapper[4733]: I0202 15:15:24.498894 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:25 crc kubenswrapper[4733]: I0202 15:15:25.193671 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 02:53:03.773445747 +0000 UTC Feb 02 15:15:25 crc kubenswrapper[4733]: I0202 15:15:25.346242 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 15:15:25 crc kubenswrapper[4733]: I0202 15:15:25.347743 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:25 crc kubenswrapper[4733]: I0202 15:15:25.347800 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:25 crc kubenswrapper[4733]: I0202 15:15:25.347821 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:26 crc kubenswrapper[4733]: I0202 15:15:26.194610 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-18 05:50:53.776117968 +0000 UTC Feb 02 15:15:27 crc kubenswrapper[4733]: I0202 15:15:27.195455 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 21:45:13.695438962 +0000 UTC Feb 02 15:15:27 crc kubenswrapper[4733]: E0202 15:15:27.335616 4733 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 02 15:15:27 crc kubenswrapper[4733]: I0202 15:15:27.495556 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Feb 02 15:15:27 crc kubenswrapper[4733]: I0202 15:15:27.495872 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 15:15:27 crc kubenswrapper[4733]: I0202 15:15:27.497866 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:27 crc kubenswrapper[4733]: I0202 15:15:27.497928 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:27 crc kubenswrapper[4733]: I0202 15:15:27.497947 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:28 crc kubenswrapper[4733]: I0202 15:15:28.196093 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-22 10:36:00.842652153 +0000 UTC Feb 02 15:15:29 crc kubenswrapper[4733]: I0202 15:15:29.184040 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 15:15:29 crc kubenswrapper[4733]: I0202 15:15:29.184328 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 15:15:29 crc kubenswrapper[4733]: I0202 15:15:29.185736 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:29 crc kubenswrapper[4733]: I0202 15:15:29.185786 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:29 crc kubenswrapper[4733]: I0202 15:15:29.185803 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:29 crc kubenswrapper[4733]: I0202 15:15:29.196544 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-14 00:18:34.312006398 +0000 UTC Feb 02 15:15:30 crc kubenswrapper[4733]: I0202 15:15:30.087620 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 15:15:30 crc kubenswrapper[4733]: I0202 15:15:30.087843 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 15:15:30 crc kubenswrapper[4733]: I0202 15:15:30.089603 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:30 crc kubenswrapper[4733]: I0202 15:15:30.089667 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:30 crc kubenswrapper[4733]: I0202 15:15:30.089685 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:30 crc kubenswrapper[4733]: I0202 15:15:30.095650 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 15:15:30 crc kubenswrapper[4733]: I0202 15:15:30.197654 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-29 09:26:56.607192662 +0000 UTC Feb 02 15:15:30 crc kubenswrapper[4733]: I0202 15:15:30.363415 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 15:15:30 crc kubenswrapper[4733]: I0202 15:15:30.364724 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:30 crc kubenswrapper[4733]: I0202 15:15:30.364791 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:30 crc kubenswrapper[4733]: I0202 15:15:30.364815 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:30 crc kubenswrapper[4733]: I0202 15:15:30.370250 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 15:15:31 crc kubenswrapper[4733]: W0202 15:15:31.147787 4733 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Feb 02 15:15:31 crc kubenswrapper[4733]: I0202 15:15:31.147897 4733 trace.go:236] Trace[806118276]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Feb-2026 15:15:21.146) (total time: 10001ms): Feb 02 15:15:31 crc kubenswrapper[4733]: Trace[806118276]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (15:15:31.147) Feb 02 15:15:31 crc kubenswrapper[4733]: Trace[806118276]: [10.001624281s] [10.001624281s] END Feb 02 15:15:31 crc kubenswrapper[4733]: E0202 15:15:31.147934 4733 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Feb 02 15:15:31 crc kubenswrapper[4733]: I0202 15:15:31.189370 4733 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Feb 02 15:15:31 crc kubenswrapper[4733]: I0202 15:15:31.197871 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-11 02:38:43.319173002 +0000 UTC Feb 02 15:15:31 crc kubenswrapper[4733]: I0202 15:15:31.365769 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 15:15:31 crc kubenswrapper[4733]: I0202 15:15:31.367360 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:31 crc kubenswrapper[4733]: I0202 15:15:31.367408 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:31 crc kubenswrapper[4733]: I0202 15:15:31.367425 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:31 crc kubenswrapper[4733]: I0202 15:15:31.421026 4733 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\": RBAC: [clusterrole.rbac.authorization.k8s.io \"system:openshift:public-info-viewer\" not found, clusterrole.rbac.authorization.k8s.io \"system:public-info-viewer\" not found]","reason":"Forbidden","details":{},"code":403} Feb 02 15:15:31 crc kubenswrapper[4733]: I0202 15:15:31.421099 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Feb 02 15:15:31 crc kubenswrapper[4733]: I0202 15:15:31.426900 4733 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\": RBAC: [clusterrole.rbac.authorization.k8s.io \"system:openshift:public-info-viewer\" not found, clusterrole.rbac.authorization.k8s.io \"system:public-info-viewer\" not found]","reason":"Forbidden","details":{},"code":403} Feb 02 15:15:31 crc kubenswrapper[4733]: I0202 15:15:31.426961 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Feb 02 15:15:32 crc kubenswrapper[4733]: I0202 15:15:32.184897 4733 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 02 15:15:32 crc kubenswrapper[4733]: I0202 15:15:32.184964 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 02 15:15:32 crc kubenswrapper[4733]: I0202 15:15:32.199022 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 09:17:21.358575612 +0000 UTC Feb 02 15:15:33 crc kubenswrapper[4733]: I0202 15:15:33.199704 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-03 14:59:34.725275134 +0000 UTC Feb 02 15:15:34 crc kubenswrapper[4733]: I0202 15:15:34.199963 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-05 09:30:23.261029205 +0000 UTC Feb 02 15:15:34 crc kubenswrapper[4733]: I0202 15:15:34.504232 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 15:15:34 crc kubenswrapper[4733]: I0202 15:15:34.504491 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 15:15:34 crc kubenswrapper[4733]: I0202 15:15:34.505969 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:34 crc kubenswrapper[4733]: I0202 15:15:34.506054 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:34 crc kubenswrapper[4733]: I0202 15:15:34.506068 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:34 crc kubenswrapper[4733]: I0202 15:15:34.511757 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 15:15:35 crc kubenswrapper[4733]: I0202 15:15:35.200763 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 00:20:09.606899334 +0000 UTC Feb 02 15:15:35 crc kubenswrapper[4733]: I0202 15:15:35.375745 4733 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 02 15:15:35 crc kubenswrapper[4733]: I0202 15:15:35.377106 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:35 crc kubenswrapper[4733]: I0202 15:15:35.377144 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:35 crc kubenswrapper[4733]: I0202 15:15:35.377155 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.201733 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-24 05:12:48.01756388 +0000 UTC Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.247324 4733 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Feb 02 15:15:36 crc kubenswrapper[4733]: E0202 15:15:36.421954 4733 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.423107 4733 trace.go:236] Trace[1998089628]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Feb-2026 15:15:21.645) (total time: 14777ms): Feb 02 15:15:36 crc kubenswrapper[4733]: Trace[1998089628]: ---"Objects listed" error: 14777ms (15:15:36.422) Feb 02 15:15:36 crc kubenswrapper[4733]: Trace[1998089628]: [14.777399258s] [14.777399258s] END Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.423153 4733 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.423598 4733 trace.go:236] Trace[836515566]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Feb-2026 15:15:26.268) (total time: 10155ms): Feb 02 15:15:36 crc kubenswrapper[4733]: Trace[836515566]: ---"Objects listed" error: 10155ms (15:15:36.423) Feb 02 15:15:36 crc kubenswrapper[4733]: Trace[836515566]: [10.155430296s] [10.155430296s] END Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.423636 4733 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.425375 4733 trace.go:236] Trace[67950237]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Feb-2026 15:15:26.401) (total time: 10023ms): Feb 02 15:15:36 crc kubenswrapper[4733]: Trace[67950237]: ---"Objects listed" error: 10023ms (15:15:36.425) Feb 02 15:15:36 crc kubenswrapper[4733]: Trace[67950237]: [10.023639637s] [10.023639637s] END Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.425414 4733 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.428576 4733 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.434473 4733 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.438153 4733 kubelet_node_status.go:115] "Node was previously registered" node="crc" Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.438589 4733 kubelet_node_status.go:79] "Successfully registered node" node="crc" Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.440528 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.440582 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.440605 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.440641 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.440681 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:36Z","lastTransitionTime":"2026-02-02T15:15:36Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.465663 4733 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:42842->192.168.126.11:17697: read: connection reset by peer" start-of-body= Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.465739 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:42842->192.168.126.11:17697: read: connection reset by peer" Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.465692 4733 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:42852->192.168.126.11:17697: read: connection reset by peer" start-of-body= Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.465817 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:42852->192.168.126.11:17697: read: connection reset by peer" Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.466151 4733 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.466261 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Feb 02 15:15:36 crc kubenswrapper[4733]: E0202 15:15:36.475558 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd742269-802e-45ec-a88e-17aee17c574e\\\",\\\"systemUUID\\\":\\\"91b64b75-3cd9-4fb9-93f0-dbd3bcdcaec4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.478842 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.478876 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.478886 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.478905 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.478917 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:36Z","lastTransitionTime":"2026-02-02T15:15:36Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.481368 4733 csr.go:261] certificate signing request csr-j42bn is approved, waiting to be issued Feb 02 15:15:36 crc kubenswrapper[4733]: E0202 15:15:36.489834 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd742269-802e-45ec-a88e-17aee17c574e\\\",\\\"systemUUID\\\":\\\"91b64b75-3cd9-4fb9-93f0-dbd3bcdcaec4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.493475 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.493529 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.493545 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.493571 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.493586 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:36Z","lastTransitionTime":"2026-02-02T15:15:36Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Feb 02 15:15:36 crc kubenswrapper[4733]: E0202 15:15:36.511914 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd742269-802e-45ec-a88e-17aee17c574e\\\",\\\"systemUUID\\\":\\\"91b64b75-3cd9-4fb9-93f0-dbd3bcdcaec4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.514452 4733 csr.go:257] certificate signing request csr-j42bn is issued Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.517551 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.517590 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.517602 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.517623 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.517634 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:36Z","lastTransitionTime":"2026-02-02T15:15:36Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Feb 02 15:15:36 crc kubenswrapper[4733]: E0202 15:15:36.534781 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd742269-802e-45ec-a88e-17aee17c574e\\\",\\\"systemUUID\\\":\\\"91b64b75-3cd9-4fb9-93f0-dbd3bcdcaec4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.542680 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.542714 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.542723 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.542738 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.542747 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:36Z","lastTransitionTime":"2026-02-02T15:15:36Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Feb 02 15:15:36 crc kubenswrapper[4733]: E0202 15:15:36.581733 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd742269-802e-45ec-a88e-17aee17c574e\\\",\\\"systemUUID\\\":\\\"91b64b75-3cd9-4fb9-93f0-dbd3bcdcaec4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:36 crc kubenswrapper[4733]: E0202 15:15:36.581895 4733 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.583488 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.583530 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.583541 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.583559 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.583575 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:36Z","lastTransitionTime":"2026-02-02T15:15:36Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.686197 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.686243 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.686256 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.686279 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.686292 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:36Z","lastTransitionTime":"2026-02-02T15:15:36Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.788583 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.788618 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.788626 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.788641 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.788649 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:36Z","lastTransitionTime":"2026-02-02T15:15:36Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.890485 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.890548 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.890569 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.890606 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.890630 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:36Z","lastTransitionTime":"2026-02-02T15:15:36Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.993461 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.993528 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.993548 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.993570 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:36 crc kubenswrapper[4733]: I0202 15:15:36.993581 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:36Z","lastTransitionTime":"2026-02-02T15:15:36Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.034421 4733 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Feb 02 15:15:37 crc kubenswrapper[4733]: W0202 15:15:37.034607 4733 reflector.go:484] k8s.io/client-go/informers/factory.go:160: watch of *v1.CSIDriver ended with: very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received Feb 02 15:15:37 crc kubenswrapper[4733]: W0202 15:15:37.034639 4733 reflector.go:484] k8s.io/client-go/informers/factory.go:160: watch of *v1.RuntimeClass ended with: very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received Feb 02 15:15:37 crc kubenswrapper[4733]: W0202 15:15:37.034723 4733 reflector.go:484] k8s.io/client-go/informers/factory.go:160: watch of *v1.Node ended with: very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received Feb 02 15:15:37 crc kubenswrapper[4733]: W0202 15:15:37.034757 4733 reflector.go:484] k8s.io/client-go/informers/factory.go:160: watch of *v1.Service ended with: very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.096899 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.096960 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.096978 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.097010 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.097080 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:37Z","lastTransitionTime":"2026-02-02T15:15:37Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.178993 4733 apiserver.go:52] "Watching apiserver" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.187969 4733 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.188552 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-nr6ps","openshift-multus/multus-jh6vk","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-ovn-kubernetes/ovnkube-node-qxvkf","openshift-dns/node-resolver-5xgpn","openshift-machine-config-operator/machine-config-daemon-zgsv9","openshift-multus/network-metrics-daemon-7r9fd","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.189015 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.189843 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.189891 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.189912 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:15:37 crc kubenswrapper[4733]: E0202 15:15:37.189940 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:15:37 crc kubenswrapper[4733]: E0202 15:15:37.190215 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.190474 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.190507 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:15:37 crc kubenswrapper[4733]: E0202 15:15:37.190597 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.190887 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-nr6ps" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.191043 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-jh6vk" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.191266 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.191427 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.191494 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-5xgpn" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.191524 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.191658 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.192140 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:15:37 crc kubenswrapper[4733]: E0202 15:15:37.192267 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.193288 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.193764 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.194087 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.195436 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.195494 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.195612 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.195614 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.196823 4733 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.198631 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.199199 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.199677 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.201070 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.201317 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.201359 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.201889 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.201330 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.202053 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.202416 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.202495 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.201599 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.202592 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.202629 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.202891 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.203020 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.203143 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-09 20:58:09.024105783 +0000 UTC Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.201915 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.203371 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:37Z","lastTransitionTime":"2026-02-02T15:15:37Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.203566 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.203655 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.203792 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.203868 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.203979 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.204139 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.205377 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.205555 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.205901 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.223697 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.232352 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.234436 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.234641 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.234720 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.234762 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.234797 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.234832 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.234867 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.234901 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.234937 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.234972 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.235006 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.235043 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.235078 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.235110 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.235144 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.235213 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.235250 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.235286 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.235321 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.235353 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.235395 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.235433 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 02 15:15:37 crc kubenswrapper[4733]: E0202 15:15:37.235456 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 15:15:37.735426143 +0000 UTC m=+21.186887511 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.235501 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.235561 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.235598 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.235603 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.235636 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.235669 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.235703 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.235737 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.235766 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.235796 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.235827 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.235857 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.235869 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.235887 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.235922 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.235953 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.235987 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.236017 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.236048 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.236078 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.236107 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.236143 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.236211 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.236213 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.236251 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.236323 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.236355 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.236386 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.236416 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.236448 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.236517 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.236546 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.236576 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.236598 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.236619 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.236605 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.236694 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.236723 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.236749 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.236744 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.236770 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.236793 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.236814 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.236835 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.236855 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.236877 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.236898 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.236924 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.236947 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.236971 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.236991 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.237012 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.237037 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.237067 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.237091 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.237113 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.237211 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.237237 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.237272 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.237294 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.237317 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.237340 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.237361 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.237382 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.237404 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.237425 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.237447 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.237468 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.237488 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.237510 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.237531 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.237555 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.237575 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.237598 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.237622 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.237643 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.237665 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.237687 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.237710 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.237731 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.237754 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.237777 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.237800 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.237822 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.237845 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.237865 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.237886 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.237909 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.237930 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.237951 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.237973 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.237996 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.238049 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.238070 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.238090 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.238114 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.238136 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.238164 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.238277 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.238301 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.238325 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.238350 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.238374 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.238434 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.238457 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.238478 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.238500 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.238521 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.238543 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.238573 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.238595 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.238617 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.238639 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.238660 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.238681 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.238704 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.238727 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.238748 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.238770 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.238800 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.238824 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.238846 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.238868 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.238890 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.238911 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.238932 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.238959 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.238980 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.239003 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.239027 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.239050 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.239071 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.239092 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.239115 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.239135 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.239158 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.239211 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.239234 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.239256 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.239281 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.239303 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.239324 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.239346 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.239370 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.239393 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.239415 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.239437 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.239459 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.239482 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.239505 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.239527 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.239552 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.239574 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.239597 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.239622 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.239644 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.239667 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.239692 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.239717 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.239741 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.239764 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.239789 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.239812 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.239833 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.239856 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.239879 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.239901 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.239926 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.239951 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.239973 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.239996 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.240019 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.240043 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.240064 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.240087 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.240112 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.240134 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.240267 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.240295 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.240357 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.240388 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.240420 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b76a05d6-d9c7-4183-ad90-b3131f225841-etc-kubernetes\") pod \"multus-jh6vk\" (UID: \"b76a05d6-d9c7-4183-ad90-b3131f225841\") " pod="openshift-multus/multus-jh6vk" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.240445 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-host-run-netns\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.240469 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f5a81ada-e393-499a-aac6-7b40349958bb-ovnkube-script-lib\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.240498 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.240525 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.240550 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.240573 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/b76a05d6-d9c7-4183-ad90-b3131f225841-host-run-multus-certs\") pod \"multus-jh6vk\" (UID: \"b76a05d6-d9c7-4183-ad90-b3131f225841\") " pod="openshift-multus/multus-jh6vk" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.240597 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b76a05d6-d9c7-4183-ad90-b3131f225841-multus-cni-dir\") pod \"multus-jh6vk\" (UID: \"b76a05d6-d9c7-4183-ad90-b3131f225841\") " pod="openshift-multus/multus-jh6vk" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.240617 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b76a05d6-d9c7-4183-ad90-b3131f225841-cnibin\") pod \"multus-jh6vk\" (UID: \"b76a05d6-d9c7-4183-ad90-b3131f225841\") " pod="openshift-multus/multus-jh6vk" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.240637 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/b76a05d6-d9c7-4183-ad90-b3131f225841-host-run-k8s-cni-cncf-io\") pod \"multus-jh6vk\" (UID: \"b76a05d6-d9c7-4183-ad90-b3131f225841\") " pod="openshift-multus/multus-jh6vk" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.240660 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-host-run-ovn-kubernetes\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.240687 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.240711 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.240733 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e43cd902-791a-4042-bc08-cf03e075be32-cni-binary-copy\") pod \"multus-additional-cni-plugins-nr6ps\" (UID: \"e43cd902-791a-4042-bc08-cf03e075be32\") " pod="openshift-multus/multus-additional-cni-plugins-nr6ps" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.240755 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/b76a05d6-d9c7-4183-ad90-b3131f225841-host-var-lib-cni-multus\") pod \"multus-jh6vk\" (UID: \"b76a05d6-d9c7-4183-ad90-b3131f225841\") " pod="openshift-multus/multus-jh6vk" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.240778 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jtscj\" (UniqueName: \"kubernetes.io/projected/d4dfd44b-b363-4162-9f43-9a8dc4a3b110-kube-api-access-jtscj\") pod \"network-metrics-daemon-7r9fd\" (UID: \"d4dfd44b-b363-4162-9f43-9a8dc4a3b110\") " pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.240829 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/83d0e810-ea5c-47aa-947b-7865b045bf94-proxy-tls\") pod \"machine-config-daemon-zgsv9\" (UID: \"83d0e810-ea5c-47aa-947b-7865b045bf94\") " pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.240848 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/b76a05d6-d9c7-4183-ad90-b3131f225841-host-var-lib-kubelet\") pod \"multus-jh6vk\" (UID: \"b76a05d6-d9c7-4183-ad90-b3131f225841\") " pod="openshift-multus/multus-jh6vk" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.240874 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/b76a05d6-d9c7-4183-ad90-b3131f225841-multus-daemon-config\") pod \"multus-jh6vk\" (UID: \"b76a05d6-d9c7-4183-ad90-b3131f225841\") " pod="openshift-multus/multus-jh6vk" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.240895 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-run-systemd\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.240916 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f5a81ada-e393-499a-aac6-7b40349958bb-ovnkube-config\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.240938 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d4dfd44b-b363-4162-9f43-9a8dc4a3b110-metrics-certs\") pod \"network-metrics-daemon-7r9fd\" (UID: \"d4dfd44b-b363-4162-9f43-9a8dc4a3b110\") " pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.240959 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b76a05d6-d9c7-4183-ad90-b3131f225841-host-var-lib-cni-bin\") pod \"multus-jh6vk\" (UID: \"b76a05d6-d9c7-4183-ad90-b3131f225841\") " pod="openshift-multus/multus-jh6vk" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.240978 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-systemd-units\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.240998 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/56681968-f264-440e-aa12-d4f81879efe9-hosts-file\") pod \"node-resolver-5xgpn\" (UID: \"56681968-f264-440e-aa12-d4f81879efe9\") " pod="openshift-dns/node-resolver-5xgpn" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.241019 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.241043 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b76a05d6-d9c7-4183-ad90-b3131f225841-multus-conf-dir\") pod \"multus-jh6vk\" (UID: \"b76a05d6-d9c7-4183-ad90-b3131f225841\") " pod="openshift-multus/multus-jh6vk" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.241065 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-host-cni-netd\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.241086 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f5a81ada-e393-499a-aac6-7b40349958bb-env-overrides\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.241107 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f5a81ada-e393-499a-aac6-7b40349958bb-ovn-node-metrics-cert\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.241127 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/83d0e810-ea5c-47aa-947b-7865b045bf94-mcd-auth-proxy-config\") pod \"machine-config-daemon-zgsv9\" (UID: \"83d0e810-ea5c-47aa-947b-7865b045bf94\") " pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.241149 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hz28\" (UniqueName: \"kubernetes.io/projected/83d0e810-ea5c-47aa-947b-7865b045bf94-kube-api-access-4hz28\") pod \"machine-config-daemon-zgsv9\" (UID: \"83d0e810-ea5c-47aa-947b-7865b045bf94\") " pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.241196 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.241222 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e43cd902-791a-4042-bc08-cf03e075be32-os-release\") pod \"multus-additional-cni-plugins-nr6ps\" (UID: \"e43cd902-791a-4042-bc08-cf03e075be32\") " pod="openshift-multus/multus-additional-cni-plugins-nr6ps" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.241246 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e43cd902-791a-4042-bc08-cf03e075be32-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-nr6ps\" (UID: \"e43cd902-791a-4042-bc08-cf03e075be32\") " pod="openshift-multus/multus-additional-cni-plugins-nr6ps" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.241270 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4z57\" (UniqueName: \"kubernetes.io/projected/f5a81ada-e393-499a-aac6-7b40349958bb-kube-api-access-r4z57\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.241291 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hv9dt\" (UniqueName: \"kubernetes.io/projected/b76a05d6-d9c7-4183-ad90-b3131f225841-kube-api-access-hv9dt\") pod \"multus-jh6vk\" (UID: \"b76a05d6-d9c7-4183-ad90-b3131f225841\") " pod="openshift-multus/multus-jh6vk" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.241311 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-etc-openvswitch\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.241333 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-run-openvswitch\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.241360 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.241380 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.241446 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b76a05d6-d9c7-4183-ad90-b3131f225841-system-cni-dir\") pod \"multus-jh6vk\" (UID: \"b76a05d6-d9c7-4183-ad90-b3131f225841\") " pod="openshift-multus/multus-jh6vk" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.241468 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b76a05d6-d9c7-4183-ad90-b3131f225841-cni-binary-copy\") pod \"multus-jh6vk\" (UID: \"b76a05d6-d9c7-4183-ad90-b3131f225841\") " pod="openshift-multus/multus-jh6vk" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.241488 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/b76a05d6-d9c7-4183-ad90-b3131f225841-hostroot\") pod \"multus-jh6vk\" (UID: \"b76a05d6-d9c7-4183-ad90-b3131f225841\") " pod="openshift-multus/multus-jh6vk" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.241509 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-host-kubelet\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.241532 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.241552 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b76a05d6-d9c7-4183-ad90-b3131f225841-host-run-netns\") pod \"multus-jh6vk\" (UID: \"b76a05d6-d9c7-4183-ad90-b3131f225841\") " pod="openshift-multus/multus-jh6vk" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.241573 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-host-slash\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.241593 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-node-log\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.241612 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-log-socket\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.241639 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdjjl\" (UniqueName: \"kubernetes.io/projected/56681968-f264-440e-aa12-d4f81879efe9-kube-api-access-qdjjl\") pod \"node-resolver-5xgpn\" (UID: \"56681968-f264-440e-aa12-d4f81879efe9\") " pod="openshift-dns/node-resolver-5xgpn" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.241662 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e43cd902-791a-4042-bc08-cf03e075be32-cnibin\") pod \"multus-additional-cni-plugins-nr6ps\" (UID: \"e43cd902-791a-4042-bc08-cf03e075be32\") " pod="openshift-multus/multus-additional-cni-plugins-nr6ps" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.241684 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfskw\" (UniqueName: \"kubernetes.io/projected/e43cd902-791a-4042-bc08-cf03e075be32-kube-api-access-mfskw\") pod \"multus-additional-cni-plugins-nr6ps\" (UID: \"e43cd902-791a-4042-bc08-cf03e075be32\") " pod="openshift-multus/multus-additional-cni-plugins-nr6ps" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.241707 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-run-ovn\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.241727 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-host-cni-bin\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.241752 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.241775 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/83d0e810-ea5c-47aa-947b-7865b045bf94-rootfs\") pod \"machine-config-daemon-zgsv9\" (UID: \"83d0e810-ea5c-47aa-947b-7865b045bf94\") " pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.241798 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b76a05d6-d9c7-4183-ad90-b3131f225841-os-release\") pod \"multus-jh6vk\" (UID: \"b76a05d6-d9c7-4183-ad90-b3131f225841\") " pod="openshift-multus/multus-jh6vk" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.241819 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/b76a05d6-d9c7-4183-ad90-b3131f225841-multus-socket-dir-parent\") pod \"multus-jh6vk\" (UID: \"b76a05d6-d9c7-4183-ad90-b3131f225841\") " pod="openshift-multus/multus-jh6vk" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.241840 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-var-lib-openvswitch\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.241866 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.241889 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.241912 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e43cd902-791a-4042-bc08-cf03e075be32-system-cni-dir\") pod \"multus-additional-cni-plugins-nr6ps\" (UID: \"e43cd902-791a-4042-bc08-cf03e075be32\") " pod="openshift-multus/multus-additional-cni-plugins-nr6ps" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.241935 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e43cd902-791a-4042-bc08-cf03e075be32-tuning-conf-dir\") pod \"multus-additional-cni-plugins-nr6ps\" (UID: \"e43cd902-791a-4042-bc08-cf03e075be32\") " pod="openshift-multus/multus-additional-cni-plugins-nr6ps" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.241985 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.242003 4733 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.242018 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.242032 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.242047 4733 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.242062 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.247345 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.248735 4733 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.252088 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.257553 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.237132 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.237683 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.237764 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.237853 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.260093 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.238083 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.238151 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.238504 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.238703 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.238794 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.239222 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.239238 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.239665 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.239743 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.260481 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.260583 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.260589 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.260615 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.239796 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.239883 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.240122 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.240334 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.240338 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.240437 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.240642 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.260882 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.241003 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.241343 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.241391 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.241482 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.241508 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.241667 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.240899 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.241912 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.242130 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.242137 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.242197 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.242252 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.243234 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.243671 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.243750 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.244058 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.244389 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.244410 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.244592 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.244780 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.244863 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.244899 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.244925 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.245147 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.245716 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.245790 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.245807 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.246088 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.246167 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.246246 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.246441 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.246477 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.246473 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.246750 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.247865 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.248024 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.248433 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.248947 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.249377 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.249423 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.249661 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.250434 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: E0202 15:15:37.250570 4733 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 02 15:15:37 crc kubenswrapper[4733]: E0202 15:15:37.261573 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-02 15:15:37.761543858 +0000 UTC m=+21.213005266 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.251077 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.252618 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.252958 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.253129 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.253374 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.253797 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.253920 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.254407 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.254437 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.254837 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.254942 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.254966 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.254997 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.257332 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.257387 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.257755 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.257971 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.258597 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.258986 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.259606 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.259780 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.259848 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.259992 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.260895 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.262000 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.262488 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.262596 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.263250 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.263364 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.263600 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: E0202 15:15:37.263721 4733 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 02 15:15:37 crc kubenswrapper[4733]: E0202 15:15:37.263783 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-02 15:15:37.763764867 +0000 UTC m=+21.215226265 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.263949 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.264058 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.264162 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.264483 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.264497 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.264511 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jh6vk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76a05d6-d9c7-4183-ad90-b3131f225841\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jh6vk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.265093 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.265104 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.265849 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.266076 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.266165 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.266246 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.266301 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.266599 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.266908 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.267024 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.267058 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.267075 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.267628 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.267690 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.267799 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.267831 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.268029 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.268131 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.268365 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.268534 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.268594 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.269002 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.270950 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.272116 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.272184 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.272629 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.273157 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.273397 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.273696 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.274145 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.274614 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.275614 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.275835 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.276088 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.276345 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.276270 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.277232 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.277664 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.277838 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.277814 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83d0e810-ea5c-47aa-947b-7865b045bf94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zgsv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.277906 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.278268 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.278373 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.278481 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.278678 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.278793 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.279348 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.279548 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.279610 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.279694 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.279831 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.279883 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.279898 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: E0202 15:15:37.279974 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 02 15:15:37 crc kubenswrapper[4733]: E0202 15:15:37.280000 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 02 15:15:37 crc kubenswrapper[4733]: E0202 15:15:37.280020 4733 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 15:15:37 crc kubenswrapper[4733]: E0202 15:15:37.280196 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-02 15:15:37.780145666 +0000 UTC m=+21.231607044 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.280376 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.280506 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.280772 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.280871 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.283034 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.283494 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.283537 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.283646 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.283853 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.283938 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.284117 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.284253 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: E0202 15:15:37.284605 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 02 15:15:37 crc kubenswrapper[4733]: E0202 15:15:37.284629 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 02 15:15:37 crc kubenswrapper[4733]: E0202 15:15:37.284644 4733 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 15:15:37 crc kubenswrapper[4733]: E0202 15:15:37.284721 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-02 15:15:37.784699756 +0000 UTC m=+21.236161124 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.285008 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.285137 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.285282 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.285318 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.285689 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.287447 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.287480 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.287557 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.288450 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.288767 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.289056 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.289270 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.289276 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5xgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56681968-f264-440e-aa12-d4f81879efe9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdjjl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5xgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.289642 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.291750 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.293373 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.293416 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.294464 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.294493 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.295820 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.296044 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.296073 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.296123 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.296251 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.298231 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7r9fd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4dfd44b-b363-4162-9f43-9a8dc4a3b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7r9fd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.298753 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.298835 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.302213 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.305659 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.306607 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.307085 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.307119 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.307131 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.307150 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.307199 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:37Z","lastTransitionTime":"2026-02-02T15:15:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.308262 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.311125 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.312206 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.312945 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.313611 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.315195 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.316032 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.317953 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.319264 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.319743 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.323980 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.324731 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.325504 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.325729 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.326433 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.327057 4733 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.327722 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.328583 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.329275 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.329598 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.330074 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.331013 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.332576 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.333240 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.334101 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.334724 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.335741 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.336199 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.337152 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.337799 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.338731 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.339206 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.339438 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.340096 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.340658 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.342063 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.342604 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/b76a05d6-d9c7-4183-ad90-b3131f225841-host-var-lib-cni-multus\") pod \"multus-jh6vk\" (UID: \"b76a05d6-d9c7-4183-ad90-b3131f225841\") " pod="openshift-multus/multus-jh6vk" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.342723 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/b76a05d6-d9c7-4183-ad90-b3131f225841-host-var-lib-cni-multus\") pod \"multus-jh6vk\" (UID: \"b76a05d6-d9c7-4183-ad90-b3131f225841\") " pod="openshift-multus/multus-jh6vk" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.342778 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.342876 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e43cd902-791a-4042-bc08-cf03e075be32-cni-binary-copy\") pod \"multus-additional-cni-plugins-nr6ps\" (UID: \"e43cd902-791a-4042-bc08-cf03e075be32\") " pod="openshift-multus/multus-additional-cni-plugins-nr6ps" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.342972 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/b76a05d6-d9c7-4183-ad90-b3131f225841-multus-daemon-config\") pod \"multus-jh6vk\" (UID: \"b76a05d6-d9c7-4183-ad90-b3131f225841\") " pod="openshift-multus/multus-jh6vk" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.343108 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-run-systemd\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.343232 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jtscj\" (UniqueName: \"kubernetes.io/projected/d4dfd44b-b363-4162-9f43-9a8dc4a3b110-kube-api-access-jtscj\") pod \"network-metrics-daemon-7r9fd\" (UID: \"d4dfd44b-b363-4162-9f43-9a8dc4a3b110\") " pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.343263 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.343275 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-run-systemd\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.343329 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/83d0e810-ea5c-47aa-947b-7865b045bf94-proxy-tls\") pod \"machine-config-daemon-zgsv9\" (UID: \"83d0e810-ea5c-47aa-947b-7865b045bf94\") " pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.343521 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/b76a05d6-d9c7-4183-ad90-b3131f225841-host-var-lib-kubelet\") pod \"multus-jh6vk\" (UID: \"b76a05d6-d9c7-4183-ad90-b3131f225841\") " pod="openshift-multus/multus-jh6vk" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.343543 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f5a81ada-e393-499a-aac6-7b40349958bb-ovnkube-config\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.343559 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d4dfd44b-b363-4162-9f43-9a8dc4a3b110-metrics-certs\") pod \"network-metrics-daemon-7r9fd\" (UID: \"d4dfd44b-b363-4162-9f43-9a8dc4a3b110\") " pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.343574 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b76a05d6-d9c7-4183-ad90-b3131f225841-host-var-lib-cni-bin\") pod \"multus-jh6vk\" (UID: \"b76a05d6-d9c7-4183-ad90-b3131f225841\") " pod="openshift-multus/multus-jh6vk" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.343574 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e43cd902-791a-4042-bc08-cf03e075be32-cni-binary-copy\") pod \"multus-additional-cni-plugins-nr6ps\" (UID: \"e43cd902-791a-4042-bc08-cf03e075be32\") " pod="openshift-multus/multus-additional-cni-plugins-nr6ps" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.343589 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-systemd-units\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.343606 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/56681968-f264-440e-aa12-d4f81879efe9-hosts-file\") pod \"node-resolver-5xgpn\" (UID: \"56681968-f264-440e-aa12-d4f81879efe9\") " pod="openshift-dns/node-resolver-5xgpn" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.343624 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.343641 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b76a05d6-d9c7-4183-ad90-b3131f225841-multus-conf-dir\") pod \"multus-jh6vk\" (UID: \"b76a05d6-d9c7-4183-ad90-b3131f225841\") " pod="openshift-multus/multus-jh6vk" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.343674 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-host-cni-netd\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.343689 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f5a81ada-e393-499a-aac6-7b40349958bb-env-overrides\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.343703 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f5a81ada-e393-499a-aac6-7b40349958bb-ovn-node-metrics-cert\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.343719 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/83d0e810-ea5c-47aa-947b-7865b045bf94-mcd-auth-proxy-config\") pod \"machine-config-daemon-zgsv9\" (UID: \"83d0e810-ea5c-47aa-947b-7865b045bf94\") " pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.343735 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hz28\" (UniqueName: \"kubernetes.io/projected/83d0e810-ea5c-47aa-947b-7865b045bf94-kube-api-access-4hz28\") pod \"machine-config-daemon-zgsv9\" (UID: \"83d0e810-ea5c-47aa-947b-7865b045bf94\") " pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.343759 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e43cd902-791a-4042-bc08-cf03e075be32-os-release\") pod \"multus-additional-cni-plugins-nr6ps\" (UID: \"e43cd902-791a-4042-bc08-cf03e075be32\") " pod="openshift-multus/multus-additional-cni-plugins-nr6ps" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.343779 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e43cd902-791a-4042-bc08-cf03e075be32-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-nr6ps\" (UID: \"e43cd902-791a-4042-bc08-cf03e075be32\") " pod="openshift-multus/multus-additional-cni-plugins-nr6ps" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.343794 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4z57\" (UniqueName: \"kubernetes.io/projected/f5a81ada-e393-499a-aac6-7b40349958bb-kube-api-access-r4z57\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.343813 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hv9dt\" (UniqueName: \"kubernetes.io/projected/b76a05d6-d9c7-4183-ad90-b3131f225841-kube-api-access-hv9dt\") pod \"multus-jh6vk\" (UID: \"b76a05d6-d9c7-4183-ad90-b3131f225841\") " pod="openshift-multus/multus-jh6vk" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.343842 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/b76a05d6-d9c7-4183-ad90-b3131f225841-host-var-lib-kubelet\") pod \"multus-jh6vk\" (UID: \"b76a05d6-d9c7-4183-ad90-b3131f225841\") " pod="openshift-multus/multus-jh6vk" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.343852 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-etc-openvswitch\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.343874 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-run-openvswitch\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.343897 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.343913 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b76a05d6-d9c7-4183-ad90-b3131f225841-system-cni-dir\") pod \"multus-jh6vk\" (UID: \"b76a05d6-d9c7-4183-ad90-b3131f225841\") " pod="openshift-multus/multus-jh6vk" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.343927 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b76a05d6-d9c7-4183-ad90-b3131f225841-cni-binary-copy\") pod \"multus-jh6vk\" (UID: \"b76a05d6-d9c7-4183-ad90-b3131f225841\") " pod="openshift-multus/multus-jh6vk" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.343942 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/b76a05d6-d9c7-4183-ad90-b3131f225841-hostroot\") pod \"multus-jh6vk\" (UID: \"b76a05d6-d9c7-4183-ad90-b3131f225841\") " pod="openshift-multus/multus-jh6vk" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.343956 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-host-cni-netd\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.343956 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-host-kubelet\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.344126 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.344142 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b76a05d6-d9c7-4183-ad90-b3131f225841-host-run-netns\") pod \"multus-jh6vk\" (UID: \"b76a05d6-d9c7-4183-ad90-b3131f225841\") " pod="openshift-multus/multus-jh6vk" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.344158 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-host-slash\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.344162 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f5a81ada-e393-499a-aac6-7b40349958bb-ovnkube-config\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.344187 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-node-log\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.344228 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-log-socket\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.344246 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdjjl\" (UniqueName: \"kubernetes.io/projected/56681968-f264-440e-aa12-d4f81879efe9-kube-api-access-qdjjl\") pod \"node-resolver-5xgpn\" (UID: \"56681968-f264-440e-aa12-d4f81879efe9\") " pod="openshift-dns/node-resolver-5xgpn" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.344263 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e43cd902-791a-4042-bc08-cf03e075be32-cnibin\") pod \"multus-additional-cni-plugins-nr6ps\" (UID: \"e43cd902-791a-4042-bc08-cf03e075be32\") " pod="openshift-multus/multus-additional-cni-plugins-nr6ps" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.344278 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfskw\" (UniqueName: \"kubernetes.io/projected/e43cd902-791a-4042-bc08-cf03e075be32-kube-api-access-mfskw\") pod \"multus-additional-cni-plugins-nr6ps\" (UID: \"e43cd902-791a-4042-bc08-cf03e075be32\") " pod="openshift-multus/multus-additional-cni-plugins-nr6ps" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.344295 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-run-ovn\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.344310 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-host-cni-bin\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.344334 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/83d0e810-ea5c-47aa-947b-7865b045bf94-rootfs\") pod \"machine-config-daemon-zgsv9\" (UID: \"83d0e810-ea5c-47aa-947b-7865b045bf94\") " pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.344388 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/b76a05d6-d9c7-4183-ad90-b3131f225841-multus-daemon-config\") pod \"multus-jh6vk\" (UID: \"b76a05d6-d9c7-4183-ad90-b3131f225841\") " pod="openshift-multus/multus-jh6vk" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.344349 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b76a05d6-d9c7-4183-ad90-b3131f225841-os-release\") pod \"multus-jh6vk\" (UID: \"b76a05d6-d9c7-4183-ad90-b3131f225841\") " pod="openshift-multus/multus-jh6vk" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.344208 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-node-log\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.344420 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-run-openvswitch\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.344468 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-log-socket\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.344516 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f5a81ada-e393-499a-aac6-7b40349958bb-env-overrides\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.344497 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b76a05d6-d9c7-4183-ad90-b3131f225841-os-release\") pod \"multus-jh6vk\" (UID: \"b76a05d6-d9c7-4183-ad90-b3131f225841\") " pod="openshift-multus/multus-jh6vk" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.344572 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-host-slash\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.344572 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/56681968-f264-440e-aa12-d4f81879efe9-hosts-file\") pod \"node-resolver-5xgpn\" (UID: \"56681968-f264-440e-aa12-d4f81879efe9\") " pod="openshift-dns/node-resolver-5xgpn" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.344589 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b76a05d6-d9c7-4183-ad90-b3131f225841-host-var-lib-cni-bin\") pod \"multus-jh6vk\" (UID: \"b76a05d6-d9c7-4183-ad90-b3131f225841\") " pod="openshift-multus/multus-jh6vk" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.344605 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/b76a05d6-d9c7-4183-ad90-b3131f225841-hostroot\") pod \"multus-jh6vk\" (UID: \"b76a05d6-d9c7-4183-ad90-b3131f225841\") " pod="openshift-multus/multus-jh6vk" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.344612 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-systemd-units\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.344620 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/b76a05d6-d9c7-4183-ad90-b3131f225841-multus-socket-dir-parent\") pod \"multus-jh6vk\" (UID: \"b76a05d6-d9c7-4183-ad90-b3131f225841\") " pod="openshift-multus/multus-jh6vk" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.344640 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.344651 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-var-lib-openvswitch\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.344672 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-etc-openvswitch\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.344737 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e43cd902-791a-4042-bc08-cf03e075be32-system-cni-dir\") pod \"multus-additional-cni-plugins-nr6ps\" (UID: \"e43cd902-791a-4042-bc08-cf03e075be32\") " pod="openshift-multus/multus-additional-cni-plugins-nr6ps" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.344794 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e43cd902-791a-4042-bc08-cf03e075be32-tuning-conf-dir\") pod \"multus-additional-cni-plugins-nr6ps\" (UID: \"e43cd902-791a-4042-bc08-cf03e075be32\") " pod="openshift-multus/multus-additional-cni-plugins-nr6ps" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.344882 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b76a05d6-d9c7-4183-ad90-b3131f225841-etc-kubernetes\") pod \"multus-jh6vk\" (UID: \"b76a05d6-d9c7-4183-ad90-b3131f225841\") " pod="openshift-multus/multus-jh6vk" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.344846 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e43cd902-791a-4042-bc08-cf03e075be32-cnibin\") pod \"multus-additional-cni-plugins-nr6ps\" (UID: \"e43cd902-791a-4042-bc08-cf03e075be32\") " pod="openshift-multus/multus-additional-cni-plugins-nr6ps" Feb 02 15:15:37 crc kubenswrapper[4733]: E0202 15:15:37.344228 4733 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.345147 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-host-run-netns\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.345241 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e43cd902-791a-4042-bc08-cf03e075be32-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-nr6ps\" (UID: \"e43cd902-791a-4042-bc08-cf03e075be32\") " pod="openshift-multus/multus-additional-cni-plugins-nr6ps" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.345289 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f5a81ada-e393-499a-aac6-7b40349958bb-ovnkube-script-lib\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.345322 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/b76a05d6-d9c7-4183-ad90-b3131f225841-multus-socket-dir-parent\") pod \"multus-jh6vk\" (UID: \"b76a05d6-d9c7-4183-ad90-b3131f225841\") " pod="openshift-multus/multus-jh6vk" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.345346 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b76a05d6-d9c7-4183-ad90-b3131f225841-etc-kubernetes\") pod \"multus-jh6vk\" (UID: \"b76a05d6-d9c7-4183-ad90-b3131f225841\") " pod="openshift-multus/multus-jh6vk" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.345400 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/b76a05d6-d9c7-4183-ad90-b3131f225841-host-run-multus-certs\") pod \"multus-jh6vk\" (UID: \"b76a05d6-d9c7-4183-ad90-b3131f225841\") " pod="openshift-multus/multus-jh6vk" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.345462 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b76a05d6-d9c7-4183-ad90-b3131f225841-multus-cni-dir\") pod \"multus-jh6vk\" (UID: \"b76a05d6-d9c7-4183-ad90-b3131f225841\") " pod="openshift-multus/multus-jh6vk" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.345490 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b76a05d6-d9c7-4183-ad90-b3131f225841-cnibin\") pod \"multus-jh6vk\" (UID: \"b76a05d6-d9c7-4183-ad90-b3131f225841\") " pod="openshift-multus/multus-jh6vk" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.345543 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/b76a05d6-d9c7-4183-ad90-b3131f225841-host-run-k8s-cni-cncf-io\") pod \"multus-jh6vk\" (UID: \"b76a05d6-d9c7-4183-ad90-b3131f225841\") " pod="openshift-multus/multus-jh6vk" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.345566 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-host-run-ovn-kubernetes\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.345708 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-host-run-ovn-kubernetes\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.346034 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.346153 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-var-lib-openvswitch\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.346263 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b76a05d6-d9c7-4183-ad90-b3131f225841-cni-binary-copy\") pod \"multus-jh6vk\" (UID: \"b76a05d6-d9c7-4183-ad90-b3131f225841\") " pod="openshift-multus/multus-jh6vk" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.346292 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b76a05d6-d9c7-4183-ad90-b3131f225841-host-run-netns\") pod \"multus-jh6vk\" (UID: \"b76a05d6-d9c7-4183-ad90-b3131f225841\") " pod="openshift-multus/multus-jh6vk" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.346347 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/83d0e810-ea5c-47aa-947b-7865b045bf94-rootfs\") pod \"machine-config-daemon-zgsv9\" (UID: \"83d0e810-ea5c-47aa-947b-7865b045bf94\") " pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.346361 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e43cd902-791a-4042-bc08-cf03e075be32-system-cni-dir\") pod \"multus-additional-cni-plugins-nr6ps\" (UID: \"e43cd902-791a-4042-bc08-cf03e075be32\") " pod="openshift-multus/multus-additional-cni-plugins-nr6ps" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.346410 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b76a05d6-d9c7-4183-ad90-b3131f225841-system-cni-dir\") pod \"multus-jh6vk\" (UID: \"b76a05d6-d9c7-4183-ad90-b3131f225841\") " pod="openshift-multus/multus-jh6vk" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.346429 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b76a05d6-d9c7-4183-ad90-b3131f225841-multus-conf-dir\") pod \"multus-jh6vk\" (UID: \"b76a05d6-d9c7-4183-ad90-b3131f225841\") " pod="openshift-multus/multus-jh6vk" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.346444 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.346469 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e43cd902-791a-4042-bc08-cf03e075be32-os-release\") pod \"multus-additional-cni-plugins-nr6ps\" (UID: \"e43cd902-791a-4042-bc08-cf03e075be32\") " pod="openshift-multus/multus-additional-cni-plugins-nr6ps" Feb 02 15:15:37 crc kubenswrapper[4733]: E0202 15:15:37.346484 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d4dfd44b-b363-4162-9f43-9a8dc4a3b110-metrics-certs podName:d4dfd44b-b363-4162-9f43-9a8dc4a3b110 nodeName:}" failed. No retries permitted until 2026-02-02 15:15:37.846463647 +0000 UTC m=+21.297925015 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d4dfd44b-b363-4162-9f43-9a8dc4a3b110-metrics-certs") pod "network-metrics-daemon-7r9fd" (UID: "d4dfd44b-b363-4162-9f43-9a8dc4a3b110") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.346488 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-host-kubelet\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.346498 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-host-cni-bin\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.346561 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-host-run-netns\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.346589 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/b76a05d6-d9c7-4183-ad90-b3131f225841-host-run-k8s-cni-cncf-io\") pod \"multus-jh6vk\" (UID: \"b76a05d6-d9c7-4183-ad90-b3131f225841\") " pod="openshift-multus/multus-jh6vk" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.346602 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-run-ovn\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.346674 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b76a05d6-d9c7-4183-ad90-b3131f225841-cnibin\") pod \"multus-jh6vk\" (UID: \"b76a05d6-d9c7-4183-ad90-b3131f225841\") " pod="openshift-multus/multus-jh6vk" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.346711 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/b76a05d6-d9c7-4183-ad90-b3131f225841-host-run-multus-certs\") pod \"multus-jh6vk\" (UID: \"b76a05d6-d9c7-4183-ad90-b3131f225841\") " pod="openshift-multus/multus-jh6vk" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.351337 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/83d0e810-ea5c-47aa-947b-7865b045bf94-mcd-auth-proxy-config\") pod \"machine-config-daemon-zgsv9\" (UID: \"83d0e810-ea5c-47aa-947b-7865b045bf94\") " pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.353016 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.353814 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f5a81ada-e393-499a-aac6-7b40349958bb-ovnkube-script-lib\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.353935 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.354295 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b76a05d6-d9c7-4183-ad90-b3131f225841-multus-cni-dir\") pod \"multus-jh6vk\" (UID: \"b76a05d6-d9c7-4183-ad90-b3131f225841\") " pod="openshift-multus/multus-jh6vk" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.354676 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.354682 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f5a81ada-e393-499a-aac6-7b40349958bb-ovn-node-metrics-cert\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.354705 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.354360 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e43cd902-791a-4042-bc08-cf03e075be32-tuning-conf-dir\") pod \"multus-additional-cni-plugins-nr6ps\" (UID: \"e43cd902-791a-4042-bc08-cf03e075be32\") " pod="openshift-multus/multus-additional-cni-plugins-nr6ps" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.354872 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.354889 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.354905 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.354916 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.354925 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.354938 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.354947 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.354957 4733 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.354967 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.354979 4733 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.354989 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.354998 4733 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355007 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355021 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355031 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355040 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355048 4733 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355059 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355069 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355079 4733 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355090 4733 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355099 4733 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355108 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355117 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355128 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355138 4733 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355148 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355157 4733 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355192 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355202 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355213 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355225 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355235 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355246 4733 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355255 4733 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355267 4733 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355276 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355287 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355297 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355311 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355320 4733 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355331 4733 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355341 4733 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355354 4733 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355363 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355373 4733 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355385 4733 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355395 4733 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355404 4733 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355416 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355429 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355439 4733 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355448 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355456 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355469 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355478 4733 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355489 4733 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355501 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355510 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355520 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355529 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355542 4733 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355551 4733 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355561 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355572 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355583 4733 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355593 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355603 4733 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355612 4733 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355623 4733 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355633 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355641 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355652 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355661 4733 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355669 4733 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355679 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355690 4733 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355700 4733 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355709 4733 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355720 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355733 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355744 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355755 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355767 4733 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355776 4733 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355785 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355798 4733 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355821 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355829 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355839 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355848 4733 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355860 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355869 4733 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355878 4733 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355887 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355899 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355908 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355917 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355929 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355938 4733 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355946 4733 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355955 4733 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355967 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355977 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355986 4733 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.355995 4733 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356007 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356015 4733 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356024 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356033 4733 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356044 4733 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356053 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356062 4733 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356074 4733 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356083 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356091 4733 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356100 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356113 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356124 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356133 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356142 4733 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356155 4733 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356164 4733 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356190 4733 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356202 4733 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356213 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356224 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356235 4733 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356248 4733 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356258 4733 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356267 4733 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356276 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356288 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356296 4733 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356306 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356314 4733 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356326 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356336 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356347 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356358 4733 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356370 4733 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356380 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356389 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356403 4733 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356411 4733 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356420 4733 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356429 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356441 4733 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356452 4733 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356461 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356471 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356482 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356491 4733 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356501 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356512 4733 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356521 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356529 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356538 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356550 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356558 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356568 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356577 4733 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356589 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356599 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356608 4733 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356619 4733 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356627 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356635 4733 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356644 4733 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356654 4733 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356662 4733 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356671 4733 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356681 4733 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356692 4733 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356702 4733 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356711 4733 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356720 4733 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356732 4733 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356741 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356752 4733 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356763 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356772 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356782 4733 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356791 4733 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356802 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.356811 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.358116 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.360145 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.360935 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nr6ps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e43cd902-791a-4042-bc08-cf03e075be32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nr6ps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.362676 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jtscj\" (UniqueName: \"kubernetes.io/projected/d4dfd44b-b363-4162-9f43-9a8dc4a3b110-kube-api-access-jtscj\") pod \"network-metrics-daemon-7r9fd\" (UID: \"d4dfd44b-b363-4162-9f43-9a8dc4a3b110\") " pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.362838 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfskw\" (UniqueName: \"kubernetes.io/projected/e43cd902-791a-4042-bc08-cf03e075be32-kube-api-access-mfskw\") pod \"multus-additional-cni-plugins-nr6ps\" (UID: \"e43cd902-791a-4042-bc08-cf03e075be32\") " pod="openshift-multus/multus-additional-cni-plugins-nr6ps" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.365914 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hz28\" (UniqueName: \"kubernetes.io/projected/83d0e810-ea5c-47aa-947b-7865b045bf94-kube-api-access-4hz28\") pod \"machine-config-daemon-zgsv9\" (UID: \"83d0e810-ea5c-47aa-947b-7865b045bf94\") " pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.367231 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4z57\" (UniqueName: \"kubernetes.io/projected/f5a81ada-e393-499a-aac6-7b40349958bb-kube-api-access-r4z57\") pod \"ovnkube-node-qxvkf\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.369045 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/83d0e810-ea5c-47aa-947b-7865b045bf94-proxy-tls\") pod \"machine-config-daemon-zgsv9\" (UID: \"83d0e810-ea5c-47aa-947b-7865b045bf94\") " pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.371064 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hv9dt\" (UniqueName: \"kubernetes.io/projected/b76a05d6-d9c7-4183-ad90-b3131f225841-kube-api-access-hv9dt\") pod \"multus-jh6vk\" (UID: \"b76a05d6-d9c7-4183-ad90-b3131f225841\") " pod="openshift-multus/multus-jh6vk" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.376021 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdjjl\" (UniqueName: \"kubernetes.io/projected/56681968-f264-440e-aa12-d4f81879efe9-kube-api-access-qdjjl\") pod \"node-resolver-5xgpn\" (UID: \"56681968-f264-440e-aa12-d4f81879efe9\") " pod="openshift-dns/node-resolver-5xgpn" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.383663 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.383587 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5a81ada-e393-499a-aac6-7b40349958bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxvkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.385542 4733 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="992f4198192cbe4b1c89d575aa4a20c29c1115fd5eaa8986c6ed70c1d6b143c4" exitCode=255 Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.385615 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"992f4198192cbe4b1c89d575aa4a20c29c1115fd5eaa8986c6ed70c1d6b143c4"} Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.391501 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83d0e810-ea5c-47aa-947b-7865b045bf94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zgsv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.401797 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.402069 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.402407 4733 scope.go:117] "RemoveContainer" containerID="992f4198192cbe4b1c89d575aa4a20c29c1115fd5eaa8986c6ed70c1d6b143c4" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.409309 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.409442 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.409538 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.409624 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.409923 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:37Z","lastTransitionTime":"2026-02-02T15:15:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.411971 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.422840 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jh6vk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76a05d6-d9c7-4183-ad90-b3131f225841\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jh6vk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.435585 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.445380 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.454693 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.462120 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5xgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56681968-f264-440e-aa12-d4f81879efe9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdjjl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5xgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.468532 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7r9fd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4dfd44b-b363-4162-9f43-9a8dc4a3b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7r9fd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.477494 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.487622 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nr6ps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e43cd902-791a-4042-bc08-cf03e075be32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nr6ps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.505916 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5a81ada-e393-499a-aac6-7b40349958bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxvkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.512862 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.512898 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.512911 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.512928 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.512940 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:37Z","lastTransitionTime":"2026-02-02T15:15:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.515518 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.516136 4733 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-02-02 15:10:36 +0000 UTC, rotation deadline is 2026-11-26 19:06:13.36201561 +0000 UTC Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.516215 4733 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 7131h50m35.845804969s for next certificate rotation Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.516225 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.517784 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.531425 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.532372 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.532494 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:37 crc kubenswrapper[4733]: W0202 15:15:37.540033 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-d2a07126bddd2f736ec1be4986ee1230c9c412fd5824a46faed42e4a6fe6f34b WatchSource:0}: Error finding container d2a07126bddd2f736ec1be4986ee1230c9c412fd5824a46faed42e4a6fe6f34b: Status 404 returned error can't find the container with id d2a07126bddd2f736ec1be4986ee1230c9c412fd5824a46faed42e4a6fe6f34b Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.548432 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jh6vk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76a05d6-d9c7-4183-ad90-b3131f225841\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jh6vk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.550197 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.565477 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83d0e810-ea5c-47aa-947b-7865b045bf94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zgsv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:37 crc kubenswrapper[4733]: W0202 15:15:37.571468 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-7c39e683c1413c367465e8135e10fde55402ede2b10be76094235fdcd338fa4f WatchSource:0}: Error finding container 7c39e683c1413c367465e8135e10fde55402ede2b10be76094235fdcd338fa4f: Status 404 returned error can't find the container with id 7c39e683c1413c367465e8135e10fde55402ede2b10be76094235fdcd338fa4f Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.580366 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7082da2-f2e7-4b73-ba35-504e5cd4c920\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd383d69b9b27d52fa8672ff6e0294a74876d676973bdc0ca95125fef5f2d085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037e379a109ce230098e3ca223fd1dea3149de3ebe601f56e2bb295f3093a3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6fdd19dac506f2d5d7eedc44ec2cda92a9381b73e91a45d949e4fcde4dfc283\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://992f4198192cbe4b1c89d575aa4a20c29c1115fd5eaa8986c6ed70c1d6b143c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://992f4198192cbe4b1c89d575aa4a20c29c1115fd5eaa8986c6ed70c1d6b143c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0202 15:15:31.010033 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 15:15:31.011243 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3543528698/tls.crt::/tmp/serving-cert-3543528698/tls.key\\\\\\\"\\\\nI0202 15:15:36.416658 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 15:15:36.428087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 15:15:36.428134 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 15:15:36.428243 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 15:15:36.428280 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 15:15:36.439678 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 15:15:36.439710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439717 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439722 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 15:15:36.439726 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 15:15:36.439729 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 15:15:36.439732 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 15:15:36.439753 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 15:15:36.450716 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://226f5949820d4f498dbf6fa251c6edf1c6356900bd946e6afae2c29042b1c835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.604419 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.610449 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-nr6ps" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.614046 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.619732 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.622172 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-jh6vk" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.625425 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.627605 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.627634 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.627642 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.627656 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.627666 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:37Z","lastTransitionTime":"2026-02-02T15:15:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.638435 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.642454 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:37 crc kubenswrapper[4733]: W0202 15:15:37.646147 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb76a05d6_d9c7_4183_ad90_b3131f225841.slice/crio-afe9c185a9f39b195d99fbf75283d14a248c58b12571d62a1bbc4dd237880093 WatchSource:0}: Error finding container afe9c185a9f39b195d99fbf75283d14a248c58b12571d62a1bbc4dd237880093: Status 404 returned error can't find the container with id afe9c185a9f39b195d99fbf75283d14a248c58b12571d62a1bbc4dd237880093 Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.652991 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5xgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56681968-f264-440e-aa12-d4f81879efe9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdjjl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5xgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.661524 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" Feb 02 15:15:37 crc kubenswrapper[4733]: W0202 15:15:37.664395 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf5a81ada_e393_499a_aac6_7b40349958bb.slice/crio-37078bebdb038f4f11bd31eb1708207108febd2a084514e18cd6e480f47faab9 WatchSource:0}: Error finding container 37078bebdb038f4f11bd31eb1708207108febd2a084514e18cd6e480f47faab9: Status 404 returned error can't find the container with id 37078bebdb038f4f11bd31eb1708207108febd2a084514e18cd6e480f47faab9 Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.665353 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7r9fd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4dfd44b-b363-4162-9f43-9a8dc4a3b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7r9fd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.674937 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nr6ps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e43cd902-791a-4042-bc08-cf03e075be32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nr6ps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.676017 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-5xgpn" Feb 02 15:15:37 crc kubenswrapper[4733]: W0202 15:15:37.690003 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod83d0e810_ea5c_47aa_947b_7865b045bf94.slice/crio-70bad89515b92b4502e43fcd27dfa527fe044590ea3ec68c620b25ee7881dcfb WatchSource:0}: Error finding container 70bad89515b92b4502e43fcd27dfa527fe044590ea3ec68c620b25ee7881dcfb: Status 404 returned error can't find the container with id 70bad89515b92b4502e43fcd27dfa527fe044590ea3ec68c620b25ee7881dcfb Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.690157 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5a81ada-e393-499a-aac6-7b40349958bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxvkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.704025 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:37 crc kubenswrapper[4733]: W0202 15:15:37.711667 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod56681968_f264_440e_aa12_d4f81879efe9.slice/crio-d1b0a070cf09d9f4648acda3e1f4ba78bdc1fddfbbda188da51522dc7b766e06 WatchSource:0}: Error finding container d1b0a070cf09d9f4648acda3e1f4ba78bdc1fddfbbda188da51522dc7b766e06: Status 404 returned error can't find the container with id d1b0a070cf09d9f4648acda3e1f4ba78bdc1fddfbbda188da51522dc7b766e06 Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.731315 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.731358 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.731368 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.731385 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.731396 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:37Z","lastTransitionTime":"2026-02-02T15:15:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.741665 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.762622 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.762719 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:15:37 crc kubenswrapper[4733]: E0202 15:15:37.762853 4733 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 02 15:15:37 crc kubenswrapper[4733]: E0202 15:15:37.762893 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-02 15:15:38.762881264 +0000 UTC m=+22.214342622 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 02 15:15:37 crc kubenswrapper[4733]: E0202 15:15:37.763134 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 15:15:38.76312711 +0000 UTC m=+22.214588458 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.775613 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jh6vk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76a05d6-d9c7-4183-ad90-b3131f225841\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jh6vk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.815429 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83d0e810-ea5c-47aa-947b-7865b045bf94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zgsv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.839343 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.839375 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.839383 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.839396 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.839405 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:37Z","lastTransitionTime":"2026-02-02T15:15:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.848502 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5xgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56681968-f264-440e-aa12-d4f81879efe9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdjjl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5xgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.867723 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.867765 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.867796 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d4dfd44b-b363-4162-9f43-9a8dc4a3b110-metrics-certs\") pod \"network-metrics-daemon-7r9fd\" (UID: \"d4dfd44b-b363-4162-9f43-9a8dc4a3b110\") " pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.867812 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:15:37 crc kubenswrapper[4733]: E0202 15:15:37.867907 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 02 15:15:37 crc kubenswrapper[4733]: E0202 15:15:37.867924 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 02 15:15:37 crc kubenswrapper[4733]: E0202 15:15:37.867935 4733 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 15:15:37 crc kubenswrapper[4733]: E0202 15:15:37.867974 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-02 15:15:38.867960191 +0000 UTC m=+22.319421539 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 15:15:37 crc kubenswrapper[4733]: E0202 15:15:37.868033 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 02 15:15:37 crc kubenswrapper[4733]: E0202 15:15:37.868046 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 02 15:15:37 crc kubenswrapper[4733]: E0202 15:15:37.868055 4733 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 15:15:37 crc kubenswrapper[4733]: E0202 15:15:37.868082 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-02 15:15:38.868074004 +0000 UTC m=+22.319535362 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 15:15:37 crc kubenswrapper[4733]: E0202 15:15:37.868137 4733 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 02 15:15:37 crc kubenswrapper[4733]: E0202 15:15:37.868207 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-02 15:15:38.868199068 +0000 UTC m=+22.319660436 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 02 15:15:37 crc kubenswrapper[4733]: E0202 15:15:37.868255 4733 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 02 15:15:37 crc kubenswrapper[4733]: E0202 15:15:37.868279 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d4dfd44b-b363-4162-9f43-9a8dc4a3b110-metrics-certs podName:d4dfd44b-b363-4162-9f43-9a8dc4a3b110 nodeName:}" failed. No retries permitted until 2026-02-02 15:15:38.868271709 +0000 UTC m=+22.319733077 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d4dfd44b-b363-4162-9f43-9a8dc4a3b110-metrics-certs") pod "network-metrics-daemon-7r9fd" (UID: "d4dfd44b-b363-4162-9f43-9a8dc4a3b110") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.886399 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7r9fd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4dfd44b-b363-4162-9f43-9a8dc4a3b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7r9fd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.928563 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7082da2-f2e7-4b73-ba35-504e5cd4c920\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd383d69b9b27d52fa8672ff6e0294a74876d676973bdc0ca95125fef5f2d085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037e379a109ce230098e3ca223fd1dea3149de3ebe601f56e2bb295f3093a3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6fdd19dac506f2d5d7eedc44ec2cda92a9381b73e91a45d949e4fcde4dfc283\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://992f4198192cbe4b1c89d575aa4a20c29c1115fd5eaa8986c6ed70c1d6b143c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://992f4198192cbe4b1c89d575aa4a20c29c1115fd5eaa8986c6ed70c1d6b143c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0202 15:15:31.010033 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 15:15:31.011243 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3543528698/tls.crt::/tmp/serving-cert-3543528698/tls.key\\\\\\\"\\\\nI0202 15:15:36.416658 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 15:15:36.428087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 15:15:36.428134 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 15:15:36.428243 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 15:15:36.428280 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 15:15:36.439678 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 15:15:36.439710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439717 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439722 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 15:15:36.439726 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 15:15:36.439729 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 15:15:36.439732 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 15:15:36.439753 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 15:15:36.450716 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://226f5949820d4f498dbf6fa251c6edf1c6356900bd946e6afae2c29042b1c835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.943146 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.943197 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.943206 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.943221 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.943231 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:37Z","lastTransitionTime":"2026-02-02T15:15:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:37 crc kubenswrapper[4733]: I0202 15:15:37.973202 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.009616 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.045042 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.045091 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.045100 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.045114 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.045123 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:38Z","lastTransitionTime":"2026-02-02T15:15:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.045636 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.086096 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.131572 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8206eca4-7d8b-4b59-a4c1-24631dd8cd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41d4a7bdc21f00b5039385d7227d8cb0f4ea3051d3b5065c71a49cb0bfad26a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c11b1859bf3bc0ec4d5e6a58d9becc1ac6333a1ed9820d83d634f1e93221c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c08b087d85be9bb83007c7e38a5de9182a5091716fb8a2ca5ded1b8191bebc40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fd4969f0f16c9aa44208d2bf578f77d9d50914a8cb1764507175aa2eb58e733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59175c8e5a7f3536168a69fd944bf3d536bcd7281eee96002343b7836a54401b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.146886 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.146937 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.146946 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.146962 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.146972 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:38Z","lastTransitionTime":"2026-02-02T15:15:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.169747 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nr6ps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e43cd902-791a-4042-bc08-cf03e075be32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nr6ps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.203958 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 23:52:03.514305398 +0000 UTC Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.212841 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5a81ada-e393-499a-aac6-7b40349958bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxvkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.249571 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.249642 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.249656 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.249674 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.249687 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:38Z","lastTransitionTime":"2026-02-02T15:15:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.352602 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.352628 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.352637 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.352649 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.352657 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:38Z","lastTransitionTime":"2026-02-02T15:15:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.389790 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-nr6ps" event={"ID":"e43cd902-791a-4042-bc08-cf03e075be32","Type":"ContainerDied","Data":"c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c"} Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.389776 4733 generic.go:334] "Generic (PLEG): container finished" podID="e43cd902-791a-4042-bc08-cf03e075be32" containerID="c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c" exitCode=0 Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.390021 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-nr6ps" event={"ID":"e43cd902-791a-4042-bc08-cf03e075be32","Type":"ContainerStarted","Data":"f848b5ff60fdd90de569f5126fa5b002608149f8f3c9e671e85930f9e2641346"} Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.391863 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"7c39e683c1413c367465e8135e10fde55402ede2b10be76094235fdcd338fa4f"} Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.396768 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jh6vk" event={"ID":"b76a05d6-d9c7-4183-ad90-b3131f225841","Type":"ContainerStarted","Data":"7f2659e2a914d3191daddbfb5ad109e487f03dabb559d324244e127add7d41d9"} Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.396804 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jh6vk" event={"ID":"b76a05d6-d9c7-4183-ad90-b3131f225841","Type":"ContainerStarted","Data":"afe9c185a9f39b195d99fbf75283d14a248c58b12571d62a1bbc4dd237880093"} Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.399496 4733 generic.go:334] "Generic (PLEG): container finished" podID="f5a81ada-e393-499a-aac6-7b40349958bb" containerID="206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf" exitCode=0 Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.399558 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" event={"ID":"f5a81ada-e393-499a-aac6-7b40349958bb","Type":"ContainerDied","Data":"206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf"} Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.399578 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" event={"ID":"f5a81ada-e393-499a-aac6-7b40349958bb","Type":"ContainerStarted","Data":"37078bebdb038f4f11bd31eb1708207108febd2a084514e18cd6e480f47faab9"} Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.402793 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"2da08f9aecc070b31df2ddc1be7c6eb2ad8ef68a931e57ca2e5c5302f4e75751"} Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.402845 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"d2a07126bddd2f736ec1be4986ee1230c9c412fd5824a46faed42e4a6fe6f34b"} Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.405212 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.407834 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"13965f5abcfb0a36c33310518955f192673e630a1aaba94b49400026de5bc2fc"} Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.408332 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.411660 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" event={"ID":"83d0e810-ea5c-47aa-947b-7865b045bf94","Type":"ContainerStarted","Data":"6144eb32f88bc106f7724e48133b568a137c4f3fbbfe30c7fcef405d3ab735a7"} Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.411705 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" event={"ID":"83d0e810-ea5c-47aa-947b-7865b045bf94","Type":"ContainerStarted","Data":"d51318ad71b42f4cc8d376aa50e44580e68da89d83d80defb996329a2ee5c2ad"} Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.411725 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" event={"ID":"83d0e810-ea5c-47aa-947b-7865b045bf94","Type":"ContainerStarted","Data":"70bad89515b92b4502e43fcd27dfa527fe044590ea3ec68c620b25ee7881dcfb"} Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.415872 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"d65b38b8410bea95465ee58803c4160039bf7c082ca47893680da7469418f429"} Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.415912 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"27003b336112760daa2079f4819641850c27fbc8415b63b5094c86195291aca8"} Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.415926 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"4722f52fa05f326307dc8ee0f2100cc788070a12a8ef8793ec0d864ab5f5fc9c"} Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.417791 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-5xgpn" event={"ID":"56681968-f264-440e-aa12-d4f81879efe9","Type":"ContainerStarted","Data":"e250891b1065a38a448aa6abd300b37543e819e87f006b03cc37983cb643982a"} Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.417832 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-5xgpn" event={"ID":"56681968-f264-440e-aa12-d4f81879efe9","Type":"ContainerStarted","Data":"d1b0a070cf09d9f4648acda3e1f4ba78bdc1fddfbbda188da51522dc7b766e06"} Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.420421 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nr6ps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e43cd902-791a-4042-bc08-cf03e075be32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nr6ps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.446108 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5a81ada-e393-499a-aac6-7b40349958bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxvkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:38Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.455603 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.455660 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.455676 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.455696 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.455710 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:38Z","lastTransitionTime":"2026-02-02T15:15:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.468656 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8206eca4-7d8b-4b59-a4c1-24631dd8cd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41d4a7bdc21f00b5039385d7227d8cb0f4ea3051d3b5065c71a49cb0bfad26a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c11b1859bf3bc0ec4d5e6a58d9becc1ac6333a1ed9820d83d634f1e93221c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c08b087d85be9bb83007c7e38a5de9182a5091716fb8a2ca5ded1b8191bebc40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fd4969f0f16c9aa44208d2bf578f77d9d50914a8cb1764507175aa2eb58e733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59175c8e5a7f3536168a69fd944bf3d536bcd7281eee96002343b7836a54401b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:38Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.482623 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:38Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.495363 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jh6vk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76a05d6-d9c7-4183-ad90-b3131f225841\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jh6vk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:38Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.505262 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83d0e810-ea5c-47aa-947b-7865b045bf94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zgsv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:38Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.517827 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:38Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.529395 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:38Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.558348 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.558377 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.558384 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.558397 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.558405 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:38Z","lastTransitionTime":"2026-02-02T15:15:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.573384 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:38Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.612311 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:38Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.663960 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.663987 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.663995 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.664007 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.664017 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:38Z","lastTransitionTime":"2026-02-02T15:15:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.669944 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:38Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.687484 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5xgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56681968-f264-440e-aa12-d4f81879efe9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdjjl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5xgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:38Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.727210 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7r9fd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4dfd44b-b363-4162-9f43-9a8dc4a3b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7r9fd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:38Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.765367 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.765404 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.765414 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.765428 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.765437 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:38Z","lastTransitionTime":"2026-02-02T15:15:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.775326 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7082da2-f2e7-4b73-ba35-504e5cd4c920\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd383d69b9b27d52fa8672ff6e0294a74876d676973bdc0ca95125fef5f2d085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037e379a109ce230098e3ca223fd1dea3149de3ebe601f56e2bb295f3093a3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6fdd19dac506f2d5d7eedc44ec2cda92a9381b73e91a45d949e4fcde4dfc283\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://992f4198192cbe4b1c89d575aa4a20c29c1115fd5eaa8986c6ed70c1d6b143c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://992f4198192cbe4b1c89d575aa4a20c29c1115fd5eaa8986c6ed70c1d6b143c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0202 15:15:31.010033 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 15:15:31.011243 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3543528698/tls.crt::/tmp/serving-cert-3543528698/tls.key\\\\\\\"\\\\nI0202 15:15:36.416658 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 15:15:36.428087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 15:15:36.428134 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 15:15:36.428243 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 15:15:36.428280 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 15:15:36.439678 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 15:15:36.439710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439717 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439722 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 15:15:36.439726 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 15:15:36.439729 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 15:15:36.439732 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 15:15:36.439753 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 15:15:36.450716 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://226f5949820d4f498dbf6fa251c6edf1c6356900bd946e6afae2c29042b1c835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:38Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.775732 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.775832 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:15:38 crc kubenswrapper[4733]: E0202 15:15:38.775960 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 15:15:40.775928396 +0000 UTC m=+24.227389754 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:15:38 crc kubenswrapper[4733]: E0202 15:15:38.775965 4733 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 02 15:15:38 crc kubenswrapper[4733]: E0202 15:15:38.776064 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-02 15:15:40.776044199 +0000 UTC m=+24.227505677 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.809798 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5xgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56681968-f264-440e-aa12-d4f81879efe9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e250891b1065a38a448aa6abd300b37543e819e87f006b03cc37983cb643982a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdjjl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5xgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:38Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.851007 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7r9fd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4dfd44b-b363-4162-9f43-9a8dc4a3b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7r9fd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:38Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.868386 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.868427 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.868437 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.868451 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.868462 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:38Z","lastTransitionTime":"2026-02-02T15:15:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.877049 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.877133 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d4dfd44b-b363-4162-9f43-9a8dc4a3b110-metrics-certs\") pod \"network-metrics-daemon-7r9fd\" (UID: \"d4dfd44b-b363-4162-9f43-9a8dc4a3b110\") " pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.877184 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.877214 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:15:38 crc kubenswrapper[4733]: E0202 15:15:38.877215 4733 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 02 15:15:38 crc kubenswrapper[4733]: E0202 15:15:38.877300 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-02 15:15:40.877282386 +0000 UTC m=+24.328743744 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 02 15:15:38 crc kubenswrapper[4733]: E0202 15:15:38.877357 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 02 15:15:38 crc kubenswrapper[4733]: E0202 15:15:38.877388 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 02 15:15:38 crc kubenswrapper[4733]: E0202 15:15:38.877401 4733 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 15:15:38 crc kubenswrapper[4733]: E0202 15:15:38.877454 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-02 15:15:40.87743623 +0000 UTC m=+24.328897678 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 15:15:38 crc kubenswrapper[4733]: E0202 15:15:38.877514 4733 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 02 15:15:38 crc kubenswrapper[4733]: E0202 15:15:38.877543 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d4dfd44b-b363-4162-9f43-9a8dc4a3b110-metrics-certs podName:d4dfd44b-b363-4162-9f43-9a8dc4a3b110 nodeName:}" failed. No retries permitted until 2026-02-02 15:15:40.877533782 +0000 UTC m=+24.328995280 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d4dfd44b-b363-4162-9f43-9a8dc4a3b110-metrics-certs") pod "network-metrics-daemon-7r9fd" (UID: "d4dfd44b-b363-4162-9f43-9a8dc4a3b110") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 02 15:15:38 crc kubenswrapper[4733]: E0202 15:15:38.877577 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 02 15:15:38 crc kubenswrapper[4733]: E0202 15:15:38.877592 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 02 15:15:38 crc kubenswrapper[4733]: E0202 15:15:38.877602 4733 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 15:15:38 crc kubenswrapper[4733]: E0202 15:15:38.877636 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-02 15:15:40.877627495 +0000 UTC m=+24.329088853 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.892956 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7082da2-f2e7-4b73-ba35-504e5cd4c920\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd383d69b9b27d52fa8672ff6e0294a74876d676973bdc0ca95125fef5f2d085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037e379a109ce230098e3ca223fd1dea3149de3ebe601f56e2bb295f3093a3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6fdd19dac506f2d5d7eedc44ec2cda92a9381b73e91a45d949e4fcde4dfc283\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13965f5abcfb0a36c33310518955f192673e630a1aaba94b49400026de5bc2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://992f4198192cbe4b1c89d575aa4a20c29c1115fd5eaa8986c6ed70c1d6b143c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0202 15:15:31.010033 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 15:15:31.011243 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3543528698/tls.crt::/tmp/serving-cert-3543528698/tls.key\\\\\\\"\\\\nI0202 15:15:36.416658 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 15:15:36.428087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 15:15:36.428134 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 15:15:36.428243 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 15:15:36.428280 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 15:15:36.439678 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 15:15:36.439710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439717 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439722 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 15:15:36.439726 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 15:15:36.439729 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 15:15:36.439732 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 15:15:36.439753 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 15:15:36.450716 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://226f5949820d4f498dbf6fa251c6edf1c6356900bd946e6afae2c29042b1c835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:38Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.933261 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da08f9aecc070b31df2ddc1be7c6eb2ad8ef68a931e57ca2e5c5302f4e75751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:38Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.969836 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:38Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.970585 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.970696 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.970759 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.970833 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:38 crc kubenswrapper[4733]: I0202 15:15:38.970901 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:38Z","lastTransitionTime":"2026-02-02T15:15:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.011538 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:39Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.049955 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:39Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.073058 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.073322 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.073356 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.073383 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.073397 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:39Z","lastTransitionTime":"2026-02-02T15:15:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.098988 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8206eca4-7d8b-4b59-a4c1-24631dd8cd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41d4a7bdc21f00b5039385d7227d8cb0f4ea3051d3b5065c71a49cb0bfad26a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c11b1859bf3bc0ec4d5e6a58d9becc1ac6333a1ed9820d83d634f1e93221c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c08b087d85be9bb83007c7e38a5de9182a5091716fb8a2ca5ded1b8191bebc40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fd4969f0f16c9aa44208d2bf578f77d9d50914a8cb1764507175aa2eb58e733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59175c8e5a7f3536168a69fd944bf3d536bcd7281eee96002343b7836a54401b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:39Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.140858 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nr6ps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e43cd902-791a-4042-bc08-cf03e075be32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nr6ps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:39Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.177410 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.177464 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.177482 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.177505 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.177521 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:39Z","lastTransitionTime":"2026-02-02T15:15:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.177530 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5a81ada-e393-499a-aac6-7b40349958bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxvkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:39Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.187089 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.191014 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.204211 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-29 13:29:48.596032807 +0000 UTC Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.211241 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:39Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.226006 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.255040 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.255106 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:15:39 crc kubenswrapper[4733]: E0202 15:15:39.255204 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.255108 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:15:39 crc kubenswrapper[4733]: E0202 15:15:39.255293 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.255378 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:15:39 crc kubenswrapper[4733]: E0202 15:15:39.255480 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:15:39 crc kubenswrapper[4733]: E0202 15:15:39.255556 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.260799 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.261442 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.262093 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.262701 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.263269 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.263770 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.264364 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.264903 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.265522 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.266016 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.266538 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.267197 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.267685 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.270393 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.270884 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.279447 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b38b8410bea95465ee58803c4160039bf7c082ca47893680da7469418f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27003b336112760daa2079f4819641850c27fbc8415b63b5094c86195291aca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:39Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.280548 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.280586 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.280595 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.280608 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.280618 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:39Z","lastTransitionTime":"2026-02-02T15:15:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.314392 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jh6vk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76a05d6-d9c7-4183-ad90-b3131f225841\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f2659e2a914d3191daddbfb5ad109e487f03dabb559d324244e127add7d41d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jh6vk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:39Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.351725 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83d0e810-ea5c-47aa-947b-7865b045bf94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6144eb32f88bc106f7724e48133b568a137c4f3fbbfe30c7fcef405d3ab735a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51318ad71b42f4cc8d376aa50e44580e68da89d83d80defb996329a2ee5c2ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zgsv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:39Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.382546 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.382579 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.382591 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.382606 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.382618 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:39Z","lastTransitionTime":"2026-02-02T15:15:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.395610 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8206eca4-7d8b-4b59-a4c1-24631dd8cd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41d4a7bdc21f00b5039385d7227d8cb0f4ea3051d3b5065c71a49cb0bfad26a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c11b1859bf3bc0ec4d5e6a58d9becc1ac6333a1ed9820d83d634f1e93221c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c08b087d85be9bb83007c7e38a5de9182a5091716fb8a2ca5ded1b8191bebc40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fd4969f0f16c9aa44208d2bf578f77d9d50914a8cb1764507175aa2eb58e733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59175c8e5a7f3536168a69fd944bf3d536bcd7281eee96002343b7836a54401b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:39Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.422985 4733 generic.go:334] "Generic (PLEG): container finished" podID="e43cd902-791a-4042-bc08-cf03e075be32" containerID="f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb" exitCode=0 Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.423081 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-nr6ps" event={"ID":"e43cd902-791a-4042-bc08-cf03e075be32","Type":"ContainerDied","Data":"f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb"} Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.429836 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" event={"ID":"f5a81ada-e393-499a-aac6-7b40349958bb","Type":"ContainerStarted","Data":"c5835cb53385359e9c58eaa5d4da2569c313e0fc265ad79a6c124e8cecb1a734"} Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.430201 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" event={"ID":"f5a81ada-e393-499a-aac6-7b40349958bb","Type":"ContainerStarted","Data":"831d85055203f153381eb51d98af0cd2e4c4a0309f837545ab3d855a680133e2"} Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.430215 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" event={"ID":"f5a81ada-e393-499a-aac6-7b40349958bb","Type":"ContainerStarted","Data":"4acf2398e2e19bd54f482cb2273c54738e61627187860bb0b6daadddc2565b63"} Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.430225 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" event={"ID":"f5a81ada-e393-499a-aac6-7b40349958bb","Type":"ContainerStarted","Data":"ffd6dc4bf5b9fd31bfbee6d9340fe981c0c5fca03a79cef9316340ff5a8b3d0c"} Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.430237 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" event={"ID":"f5a81ada-e393-499a-aac6-7b40349958bb","Type":"ContainerStarted","Data":"db1db75e392ebea8b387b85cabc4276d553bf2c06694cf7f294a383909f32826"} Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.439627 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nr6ps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e43cd902-791a-4042-bc08-cf03e075be32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nr6ps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:39Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.475337 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5a81ada-e393-499a-aac6-7b40349958bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxvkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:39Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.485015 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.485060 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.485072 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.485090 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.485102 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:39Z","lastTransitionTime":"2026-02-02T15:15:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.514682 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"402a6dc5-6bc7-4d7c-a909-9bb4ca48365d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d5b032497b592363a03f89cb31ba1b551f479004072dddffce540c41190ab81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a3fd02d55a10647428a05a9f53a1429a9827059c523d5deab3c4581c8f1b933\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11600ada13fbc7db1831de385202bc1f041efb656cc58aa49b0f71fe902db9bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fd39f5a05dec7a756dfd519f5df3ef7c77bbb5b8633c5af18a09ff8029cb079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:39Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.554619 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:39Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.587251 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.587293 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.587304 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.587323 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.587337 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:39Z","lastTransitionTime":"2026-02-02T15:15:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.589569 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b38b8410bea95465ee58803c4160039bf7c082ca47893680da7469418f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27003b336112760daa2079f4819641850c27fbc8415b63b5094c86195291aca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:39Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.630882 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jh6vk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76a05d6-d9c7-4183-ad90-b3131f225841\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f2659e2a914d3191daddbfb5ad109e487f03dabb559d324244e127add7d41d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jh6vk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:39Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.669411 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83d0e810-ea5c-47aa-947b-7865b045bf94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6144eb32f88bc106f7724e48133b568a137c4f3fbbfe30c7fcef405d3ab735a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51318ad71b42f4cc8d376aa50e44580e68da89d83d80defb996329a2ee5c2ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zgsv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:39Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.689277 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.689303 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.689311 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.689326 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.689335 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:39Z","lastTransitionTime":"2026-02-02T15:15:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.708380 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7r9fd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4dfd44b-b363-4162-9f43-9a8dc4a3b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7r9fd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:39Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.750587 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7082da2-f2e7-4b73-ba35-504e5cd4c920\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd383d69b9b27d52fa8672ff6e0294a74876d676973bdc0ca95125fef5f2d085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037e379a109ce230098e3ca223fd1dea3149de3ebe601f56e2bb295f3093a3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6fdd19dac506f2d5d7eedc44ec2cda92a9381b73e91a45d949e4fcde4dfc283\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13965f5abcfb0a36c33310518955f192673e630a1aaba94b49400026de5bc2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://992f4198192cbe4b1c89d575aa4a20c29c1115fd5eaa8986c6ed70c1d6b143c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0202 15:15:31.010033 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 15:15:31.011243 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3543528698/tls.crt::/tmp/serving-cert-3543528698/tls.key\\\\\\\"\\\\nI0202 15:15:36.416658 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 15:15:36.428087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 15:15:36.428134 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 15:15:36.428243 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 15:15:36.428280 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 15:15:36.439678 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 15:15:36.439710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439717 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439722 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 15:15:36.439726 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 15:15:36.439729 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 15:15:36.439732 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 15:15:36.439753 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 15:15:36.450716 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://226f5949820d4f498dbf6fa251c6edf1c6356900bd946e6afae2c29042b1c835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:39Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.791103 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.791151 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.791262 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.791278 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.791287 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:39Z","lastTransitionTime":"2026-02-02T15:15:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.791755 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da08f9aecc070b31df2ddc1be7c6eb2ad8ef68a931e57ca2e5c5302f4e75751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:39Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.833262 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:39Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.870449 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:39Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.892999 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.893039 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.893048 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.893061 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.893073 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:39Z","lastTransitionTime":"2026-02-02T15:15:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.908787 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:39Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.948619 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5xgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56681968-f264-440e-aa12-d4f81879efe9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e250891b1065a38a448aa6abd300b37543e819e87f006b03cc37983cb643982a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdjjl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5xgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:39Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.989876 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da08f9aecc070b31df2ddc1be7c6eb2ad8ef68a931e57ca2e5c5302f4e75751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:39Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.995316 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.995350 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.995362 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.995376 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:39 crc kubenswrapper[4733]: I0202 15:15:39.995387 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:39Z","lastTransitionTime":"2026-02-02T15:15:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.032426 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:40Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.067753 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:40Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.084189 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-8c597"] Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.084666 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-8c597" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.097685 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.097719 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.097729 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.097813 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.097832 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:40Z","lastTransitionTime":"2026-02-02T15:15:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.109667 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:40Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.121342 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.141528 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.160656 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.181585 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.188795 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwr7v\" (UniqueName: \"kubernetes.io/projected/76cc68c3-9e90-45de-87b2-6ea1b68eadce-kube-api-access-qwr7v\") pod \"node-ca-8c597\" (UID: \"76cc68c3-9e90-45de-87b2-6ea1b68eadce\") " pod="openshift-image-registry/node-ca-8c597" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.188825 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/76cc68c3-9e90-45de-87b2-6ea1b68eadce-serviceca\") pod \"node-ca-8c597\" (UID: \"76cc68c3-9e90-45de-87b2-6ea1b68eadce\") " pod="openshift-image-registry/node-ca-8c597" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.188853 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/76cc68c3-9e90-45de-87b2-6ea1b68eadce-host\") pod \"node-ca-8c597\" (UID: \"76cc68c3-9e90-45de-87b2-6ea1b68eadce\") " pod="openshift-image-registry/node-ca-8c597" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.200274 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.200328 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.200340 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.200356 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.200367 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:40Z","lastTransitionTime":"2026-02-02T15:15:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.204298 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-28 14:01:55.670413039 +0000 UTC Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.228346 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5xgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56681968-f264-440e-aa12-d4f81879efe9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e250891b1065a38a448aa6abd300b37543e819e87f006b03cc37983cb643982a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdjjl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5xgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:40Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.267143 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7r9fd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4dfd44b-b363-4162-9f43-9a8dc4a3b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7r9fd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:40Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.290364 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwr7v\" (UniqueName: \"kubernetes.io/projected/76cc68c3-9e90-45de-87b2-6ea1b68eadce-kube-api-access-qwr7v\") pod \"node-ca-8c597\" (UID: \"76cc68c3-9e90-45de-87b2-6ea1b68eadce\") " pod="openshift-image-registry/node-ca-8c597" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.290435 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/76cc68c3-9e90-45de-87b2-6ea1b68eadce-serviceca\") pod \"node-ca-8c597\" (UID: \"76cc68c3-9e90-45de-87b2-6ea1b68eadce\") " pod="openshift-image-registry/node-ca-8c597" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.290470 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/76cc68c3-9e90-45de-87b2-6ea1b68eadce-host\") pod \"node-ca-8c597\" (UID: \"76cc68c3-9e90-45de-87b2-6ea1b68eadce\") " pod="openshift-image-registry/node-ca-8c597" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.290611 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/76cc68c3-9e90-45de-87b2-6ea1b68eadce-host\") pod \"node-ca-8c597\" (UID: \"76cc68c3-9e90-45de-87b2-6ea1b68eadce\") " pod="openshift-image-registry/node-ca-8c597" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.291434 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/76cc68c3-9e90-45de-87b2-6ea1b68eadce-serviceca\") pod \"node-ca-8c597\" (UID: \"76cc68c3-9e90-45de-87b2-6ea1b68eadce\") " pod="openshift-image-registry/node-ca-8c597" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.302180 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.302203 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.302212 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.302225 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.302237 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:40Z","lastTransitionTime":"2026-02-02T15:15:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.309631 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7082da2-f2e7-4b73-ba35-504e5cd4c920\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd383d69b9b27d52fa8672ff6e0294a74876d676973bdc0ca95125fef5f2d085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037e379a109ce230098e3ca223fd1dea3149de3ebe601f56e2bb295f3093a3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6fdd19dac506f2d5d7eedc44ec2cda92a9381b73e91a45d949e4fcde4dfc283\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13965f5abcfb0a36c33310518955f192673e630a1aaba94b49400026de5bc2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://992f4198192cbe4b1c89d575aa4a20c29c1115fd5eaa8986c6ed70c1d6b143c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0202 15:15:31.010033 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 15:15:31.011243 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3543528698/tls.crt::/tmp/serving-cert-3543528698/tls.key\\\\\\\"\\\\nI0202 15:15:36.416658 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 15:15:36.428087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 15:15:36.428134 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 15:15:36.428243 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 15:15:36.428280 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 15:15:36.439678 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 15:15:36.439710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439717 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439722 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 15:15:36.439726 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 15:15:36.439729 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 15:15:36.439732 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 15:15:36.439753 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 15:15:36.450716 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://226f5949820d4f498dbf6fa251c6edf1c6356900bd946e6afae2c29042b1c835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:40Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.339399 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwr7v\" (UniqueName: \"kubernetes.io/projected/76cc68c3-9e90-45de-87b2-6ea1b68eadce-kube-api-access-qwr7v\") pod \"node-ca-8c597\" (UID: \"76cc68c3-9e90-45de-87b2-6ea1b68eadce\") " pod="openshift-image-registry/node-ca-8c597" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.373667 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nr6ps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e43cd902-791a-4042-bc08-cf03e075be32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nr6ps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:40Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.401605 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-8c597" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.404029 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.404063 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.404075 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.404092 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.404104 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:40Z","lastTransitionTime":"2026-02-02T15:15:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.418147 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5a81ada-e393-499a-aac6-7b40349958bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxvkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:40Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.431461 4733 generic.go:334] "Generic (PLEG): container finished" podID="e43cd902-791a-4042-bc08-cf03e075be32" containerID="019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347" exitCode=0 Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.431529 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-nr6ps" event={"ID":"e43cd902-791a-4042-bc08-cf03e075be32","Type":"ContainerDied","Data":"019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347"} Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.433793 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"ca57dabf1d1ee5c6fc29291b1be687ca7e97c72650084b02067057fccac41a1c"} Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.439890 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" event={"ID":"f5a81ada-e393-499a-aac6-7b40349958bb","Type":"ContainerStarted","Data":"1380122d1694c1286f1102e6396c6504b91f63c6eb157689dfc0ea7188c17f68"} Feb 02 15:15:40 crc kubenswrapper[4733]: W0202 15:15:40.452867 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod76cc68c3_9e90_45de_87b2_6ea1b68eadce.slice/crio-ea238ef8fb7d21bd9b28a072b8edfb2155f5467a40287f2cec12b5fa26425111 WatchSource:0}: Error finding container ea238ef8fb7d21bd9b28a072b8edfb2155f5467a40287f2cec12b5fa26425111: Status 404 returned error can't find the container with id ea238ef8fb7d21bd9b28a072b8edfb2155f5467a40287f2cec12b5fa26425111 Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.458329 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8206eca4-7d8b-4b59-a4c1-24631dd8cd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41d4a7bdc21f00b5039385d7227d8cb0f4ea3051d3b5065c71a49cb0bfad26a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c11b1859bf3bc0ec4d5e6a58d9becc1ac6333a1ed9820d83d634f1e93221c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c08b087d85be9bb83007c7e38a5de9182a5091716fb8a2ca5ded1b8191bebc40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fd4969f0f16c9aa44208d2bf578f77d9d50914a8cb1764507175aa2eb58e733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59175c8e5a7f3536168a69fd944bf3d536bcd7281eee96002343b7836a54401b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:40Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.496375 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"402a6dc5-6bc7-4d7c-a909-9bb4ca48365d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d5b032497b592363a03f89cb31ba1b551f479004072dddffce540c41190ab81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a3fd02d55a10647428a05a9f53a1429a9827059c523d5deab3c4581c8f1b933\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11600ada13fbc7db1831de385202bc1f041efb656cc58aa49b0f71fe902db9bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fd39f5a05dec7a756dfd519f5df3ef7c77bbb5b8633c5af18a09ff8029cb079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:40Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.507668 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.507722 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.507732 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.507745 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.507756 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:40Z","lastTransitionTime":"2026-02-02T15:15:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.529339 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b38b8410bea95465ee58803c4160039bf7c082ca47893680da7469418f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27003b336112760daa2079f4819641850c27fbc8415b63b5094c86195291aca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:40Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.571281 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jh6vk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76a05d6-d9c7-4183-ad90-b3131f225841\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f2659e2a914d3191daddbfb5ad109e487f03dabb559d324244e127add7d41d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jh6vk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:40Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.610064 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.610108 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.610123 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.610139 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.610150 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:40Z","lastTransitionTime":"2026-02-02T15:15:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.611519 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83d0e810-ea5c-47aa-947b-7865b045bf94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6144eb32f88bc106f7724e48133b568a137c4f3fbbfe30c7fcef405d3ab735a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51318ad71b42f4cc8d376aa50e44580e68da89d83d80defb996329a2ee5c2ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zgsv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:40Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.650416 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:40Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.689026 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:40Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.713275 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.713341 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.713355 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.713375 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.713389 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:40Z","lastTransitionTime":"2026-02-02T15:15:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.730050 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b38b8410bea95465ee58803c4160039bf7c082ca47893680da7469418f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27003b336112760daa2079f4819641850c27fbc8415b63b5094c86195291aca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:40Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.771088 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jh6vk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76a05d6-d9c7-4183-ad90-b3131f225841\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f2659e2a914d3191daddbfb5ad109e487f03dabb559d324244e127add7d41d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jh6vk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:40Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.781540 4733 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.794978 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 15:15:40 crc kubenswrapper[4733]: E0202 15:15:40.795182 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 15:15:44.795141522 +0000 UTC m=+28.246602880 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.795285 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:15:40 crc kubenswrapper[4733]: E0202 15:15:40.795471 4733 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 02 15:15:40 crc kubenswrapper[4733]: E0202 15:15:40.795593 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-02 15:15:44.795566994 +0000 UTC m=+28.247028392 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.816087 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.816143 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.816192 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.816221 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.816239 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:40Z","lastTransitionTime":"2026-02-02T15:15:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.828093 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83d0e810-ea5c-47aa-947b-7865b045bf94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6144eb32f88bc106f7724e48133b568a137c4f3fbbfe30c7fcef405d3ab735a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51318ad71b42f4cc8d376aa50e44580e68da89d83d80defb996329a2ee5c2ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zgsv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:40Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.869770 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca57dabf1d1ee5c6fc29291b1be687ca7e97c72650084b02067057fccac41a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:40Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.896763 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d4dfd44b-b363-4162-9f43-9a8dc4a3b110-metrics-certs\") pod \"network-metrics-daemon-7r9fd\" (UID: \"d4dfd44b-b363-4162-9f43-9a8dc4a3b110\") " pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.896832 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.896874 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.896919 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:15:40 crc kubenswrapper[4733]: E0202 15:15:40.896955 4733 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 02 15:15:40 crc kubenswrapper[4733]: E0202 15:15:40.897051 4733 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 02 15:15:40 crc kubenswrapper[4733]: E0202 15:15:40.897065 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d4dfd44b-b363-4162-9f43-9a8dc4a3b110-metrics-certs podName:d4dfd44b-b363-4162-9f43-9a8dc4a3b110 nodeName:}" failed. No retries permitted until 2026-02-02 15:15:44.897037876 +0000 UTC m=+28.348499264 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d4dfd44b-b363-4162-9f43-9a8dc4a3b110-metrics-certs") pod "network-metrics-daemon-7r9fd" (UID: "d4dfd44b-b363-4162-9f43-9a8dc4a3b110") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 02 15:15:40 crc kubenswrapper[4733]: E0202 15:15:40.897114 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-02 15:15:44.897093508 +0000 UTC m=+28.348554906 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 02 15:15:40 crc kubenswrapper[4733]: E0202 15:15:40.897119 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 02 15:15:40 crc kubenswrapper[4733]: E0202 15:15:40.897205 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 02 15:15:40 crc kubenswrapper[4733]: E0202 15:15:40.897226 4733 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 15:15:40 crc kubenswrapper[4733]: E0202 15:15:40.897299 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-02 15:15:44.897275392 +0000 UTC m=+28.348736791 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 15:15:40 crc kubenswrapper[4733]: E0202 15:15:40.897128 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 02 15:15:40 crc kubenswrapper[4733]: E0202 15:15:40.897359 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 02 15:15:40 crc kubenswrapper[4733]: E0202 15:15:40.897374 4733 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 15:15:40 crc kubenswrapper[4733]: E0202 15:15:40.897418 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-02 15:15:44.897405296 +0000 UTC m=+28.348866694 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.914945 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:40Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.918651 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.918705 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.918724 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.918750 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.918773 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:40Z","lastTransitionTime":"2026-02-02T15:15:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.949255 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5xgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56681968-f264-440e-aa12-d4f81879efe9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e250891b1065a38a448aa6abd300b37543e819e87f006b03cc37983cb643982a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdjjl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5xgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:40Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:40 crc kubenswrapper[4733]: I0202 15:15:40.992992 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7r9fd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4dfd44b-b363-4162-9f43-9a8dc4a3b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7r9fd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:40Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.021658 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.021696 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.021708 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.021726 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.021741 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:41Z","lastTransitionTime":"2026-02-02T15:15:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.034258 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7082da2-f2e7-4b73-ba35-504e5cd4c920\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd383d69b9b27d52fa8672ff6e0294a74876d676973bdc0ca95125fef5f2d085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037e379a109ce230098e3ca223fd1dea3149de3ebe601f56e2bb295f3093a3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6fdd19dac506f2d5d7eedc44ec2cda92a9381b73e91a45d949e4fcde4dfc283\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13965f5abcfb0a36c33310518955f192673e630a1aaba94b49400026de5bc2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://992f4198192cbe4b1c89d575aa4a20c29c1115fd5eaa8986c6ed70c1d6b143c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0202 15:15:31.010033 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 15:15:31.011243 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3543528698/tls.crt::/tmp/serving-cert-3543528698/tls.key\\\\\\\"\\\\nI0202 15:15:36.416658 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 15:15:36.428087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 15:15:36.428134 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 15:15:36.428243 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 15:15:36.428280 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 15:15:36.439678 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 15:15:36.439710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439717 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439722 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 15:15:36.439726 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 15:15:36.439729 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 15:15:36.439732 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 15:15:36.439753 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 15:15:36.450716 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://226f5949820d4f498dbf6fa251c6edf1c6356900bd946e6afae2c29042b1c835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:41Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.075139 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da08f9aecc070b31df2ddc1be7c6eb2ad8ef68a931e57ca2e5c5302f4e75751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:41Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.114471 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:41Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.124820 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.124871 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.124888 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.124911 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.124927 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:41Z","lastTransitionTime":"2026-02-02T15:15:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.171649 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8206eca4-7d8b-4b59-a4c1-24631dd8cd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41d4a7bdc21f00b5039385d7227d8cb0f4ea3051d3b5065c71a49cb0bfad26a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c11b1859bf3bc0ec4d5e6a58d9becc1ac6333a1ed9820d83d634f1e93221c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c08b087d85be9bb83007c7e38a5de9182a5091716fb8a2ca5ded1b8191bebc40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fd4969f0f16c9aa44208d2bf578f77d9d50914a8cb1764507175aa2eb58e733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59175c8e5a7f3536168a69fd944bf3d536bcd7281eee96002343b7836a54401b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:41Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.196936 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nr6ps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e43cd902-791a-4042-bc08-cf03e075be32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nr6ps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:41Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.205349 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-08 22:49:36.9435082 +0000 UTC Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.231305 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.231344 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.231356 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.231379 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.231398 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:41Z","lastTransitionTime":"2026-02-02T15:15:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.244597 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5a81ada-e393-499a-aac6-7b40349958bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxvkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:41Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.254957 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.255042 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:15:41 crc kubenswrapper[4733]: E0202 15:15:41.255115 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:15:41 crc kubenswrapper[4733]: E0202 15:15:41.255435 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.255561 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:15:41 crc kubenswrapper[4733]: E0202 15:15:41.255681 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.255750 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:15:41 crc kubenswrapper[4733]: E0202 15:15:41.255836 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.274003 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8c597" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76cc68c3-9e90-45de-87b2-6ea1b68eadce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwr7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8c597\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:41Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.310710 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"402a6dc5-6bc7-4d7c-a909-9bb4ca48365d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d5b032497b592363a03f89cb31ba1b551f479004072dddffce540c41190ab81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a3fd02d55a10647428a05a9f53a1429a9827059c523d5deab3c4581c8f1b933\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11600ada13fbc7db1831de385202bc1f041efb656cc58aa49b0f71fe902db9bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fd39f5a05dec7a756dfd519f5df3ef7c77bbb5b8633c5af18a09ff8029cb079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:41Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.334057 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.334116 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.334134 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.334159 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.334219 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:41Z","lastTransitionTime":"2026-02-02T15:15:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.436946 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.437021 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.437040 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.437078 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.437120 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:41Z","lastTransitionTime":"2026-02-02T15:15:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.449121 4733 generic.go:334] "Generic (PLEG): container finished" podID="e43cd902-791a-4042-bc08-cf03e075be32" containerID="52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722" exitCode=0 Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.449214 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-nr6ps" event={"ID":"e43cd902-791a-4042-bc08-cf03e075be32","Type":"ContainerDied","Data":"52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722"} Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.452998 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-8c597" event={"ID":"76cc68c3-9e90-45de-87b2-6ea1b68eadce","Type":"ContainerStarted","Data":"4ac57831c1c880e64aaa7bedcf999133ee09fc5ecc5124b9d54ecedcc9a13c28"} Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.453067 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-8c597" event={"ID":"76cc68c3-9e90-45de-87b2-6ea1b68eadce","Type":"ContainerStarted","Data":"ea238ef8fb7d21bd9b28a072b8edfb2155f5467a40287f2cec12b5fa26425111"} Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.475775 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nr6ps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e43cd902-791a-4042-bc08-cf03e075be32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nr6ps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:41Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.494562 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5a81ada-e393-499a-aac6-7b40349958bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxvkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:41Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.508537 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8c597" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76cc68c3-9e90-45de-87b2-6ea1b68eadce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwr7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8c597\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:41Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.538063 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8206eca4-7d8b-4b59-a4c1-24631dd8cd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41d4a7bdc21f00b5039385d7227d8cb0f4ea3051d3b5065c71a49cb0bfad26a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c11b1859bf3bc0ec4d5e6a58d9becc1ac6333a1ed9820d83d634f1e93221c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c08b087d85be9bb83007c7e38a5de9182a5091716fb8a2ca5ded1b8191bebc40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fd4969f0f16c9aa44208d2bf578f77d9d50914a8cb1764507175aa2eb58e733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59175c8e5a7f3536168a69fd944bf3d536bcd7281eee96002343b7836a54401b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:41Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.539814 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.539866 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.539878 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.539895 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.539909 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:41Z","lastTransitionTime":"2026-02-02T15:15:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.558833 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"402a6dc5-6bc7-4d7c-a909-9bb4ca48365d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d5b032497b592363a03f89cb31ba1b551f479004072dddffce540c41190ab81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a3fd02d55a10647428a05a9f53a1429a9827059c523d5deab3c4581c8f1b933\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11600ada13fbc7db1831de385202bc1f041efb656cc58aa49b0f71fe902db9bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fd39f5a05dec7a756dfd519f5df3ef7c77bbb5b8633c5af18a09ff8029cb079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:41Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.576844 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b38b8410bea95465ee58803c4160039bf7c082ca47893680da7469418f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27003b336112760daa2079f4819641850c27fbc8415b63b5094c86195291aca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:41Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.593794 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jh6vk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76a05d6-d9c7-4183-ad90-b3131f225841\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f2659e2a914d3191daddbfb5ad109e487f03dabb559d324244e127add7d41d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jh6vk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:41Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.633109 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83d0e810-ea5c-47aa-947b-7865b045bf94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6144eb32f88bc106f7724e48133b568a137c4f3fbbfe30c7fcef405d3ab735a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51318ad71b42f4cc8d376aa50e44580e68da89d83d80defb996329a2ee5c2ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zgsv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:41Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.644850 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.644941 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.644954 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.644972 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.645012 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:41Z","lastTransitionTime":"2026-02-02T15:15:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.677185 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:41Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.715363 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da08f9aecc070b31df2ddc1be7c6eb2ad8ef68a931e57ca2e5c5302f4e75751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:41Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.747775 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.747875 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.747894 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.747920 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.747942 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:41Z","lastTransitionTime":"2026-02-02T15:15:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.753107 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:41Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.794694 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca57dabf1d1ee5c6fc29291b1be687ca7e97c72650084b02067057fccac41a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:41Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.836647 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:41Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.851105 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.851145 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.851156 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.851215 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.851229 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:41Z","lastTransitionTime":"2026-02-02T15:15:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.873467 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5xgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56681968-f264-440e-aa12-d4f81879efe9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e250891b1065a38a448aa6abd300b37543e819e87f006b03cc37983cb643982a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdjjl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5xgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:41Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.915322 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7r9fd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4dfd44b-b363-4162-9f43-9a8dc4a3b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7r9fd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:41Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.954079 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.954138 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.954189 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.954223 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.954247 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:41Z","lastTransitionTime":"2026-02-02T15:15:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.956442 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7082da2-f2e7-4b73-ba35-504e5cd4c920\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd383d69b9b27d52fa8672ff6e0294a74876d676973bdc0ca95125fef5f2d085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037e379a109ce230098e3ca223fd1dea3149de3ebe601f56e2bb295f3093a3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6fdd19dac506f2d5d7eedc44ec2cda92a9381b73e91a45d949e4fcde4dfc283\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13965f5abcfb0a36c33310518955f192673e630a1aaba94b49400026de5bc2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://992f4198192cbe4b1c89d575aa4a20c29c1115fd5eaa8986c6ed70c1d6b143c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0202 15:15:31.010033 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 15:15:31.011243 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3543528698/tls.crt::/tmp/serving-cert-3543528698/tls.key\\\\\\\"\\\\nI0202 15:15:36.416658 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 15:15:36.428087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 15:15:36.428134 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 15:15:36.428243 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 15:15:36.428280 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 15:15:36.439678 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 15:15:36.439710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439717 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439722 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 15:15:36.439726 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 15:15:36.439729 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 15:15:36.439732 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 15:15:36.439753 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 15:15:36.450716 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://226f5949820d4f498dbf6fa251c6edf1c6356900bd946e6afae2c29042b1c835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:41Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:41 crc kubenswrapper[4733]: I0202 15:15:41.998104 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"402a6dc5-6bc7-4d7c-a909-9bb4ca48365d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d5b032497b592363a03f89cb31ba1b551f479004072dddffce540c41190ab81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a3fd02d55a10647428a05a9f53a1429a9827059c523d5deab3c4581c8f1b933\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11600ada13fbc7db1831de385202bc1f041efb656cc58aa49b0f71fe902db9bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fd39f5a05dec7a756dfd519f5df3ef7c77bbb5b8633c5af18a09ff8029cb079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:41Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.034154 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83d0e810-ea5c-47aa-947b-7865b045bf94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6144eb32f88bc106f7724e48133b568a137c4f3fbbfe30c7fcef405d3ab735a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51318ad71b42f4cc8d376aa50e44580e68da89d83d80defb996329a2ee5c2ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zgsv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:42Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.057197 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.057246 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.057259 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.057276 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.057290 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:42Z","lastTransitionTime":"2026-02-02T15:15:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.075247 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:42Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.118630 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b38b8410bea95465ee58803c4160039bf7c082ca47893680da7469418f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27003b336112760daa2079f4819641850c27fbc8415b63b5094c86195291aca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:42Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.154823 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jh6vk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76a05d6-d9c7-4183-ad90-b3131f225841\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f2659e2a914d3191daddbfb5ad109e487f03dabb559d324244e127add7d41d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jh6vk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:42Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.160035 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.160071 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.160087 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.160108 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.160123 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:42Z","lastTransitionTime":"2026-02-02T15:15:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.196654 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:42Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.205521 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 15:35:36.542085676 +0000 UTC Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.234403 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca57dabf1d1ee5c6fc29291b1be687ca7e97c72650084b02067057fccac41a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:42Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.263395 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.263457 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.263475 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.263499 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.263519 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:42Z","lastTransitionTime":"2026-02-02T15:15:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.276577 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:42Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.311410 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5xgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56681968-f264-440e-aa12-d4f81879efe9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e250891b1065a38a448aa6abd300b37543e819e87f006b03cc37983cb643982a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdjjl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5xgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:42Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.354219 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7r9fd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4dfd44b-b363-4162-9f43-9a8dc4a3b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7r9fd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:42Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.366013 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.366062 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.366078 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.366106 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.366124 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:42Z","lastTransitionTime":"2026-02-02T15:15:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.392818 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7082da2-f2e7-4b73-ba35-504e5cd4c920\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd383d69b9b27d52fa8672ff6e0294a74876d676973bdc0ca95125fef5f2d085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037e379a109ce230098e3ca223fd1dea3149de3ebe601f56e2bb295f3093a3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6fdd19dac506f2d5d7eedc44ec2cda92a9381b73e91a45d949e4fcde4dfc283\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13965f5abcfb0a36c33310518955f192673e630a1aaba94b49400026de5bc2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://992f4198192cbe4b1c89d575aa4a20c29c1115fd5eaa8986c6ed70c1d6b143c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0202 15:15:31.010033 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 15:15:31.011243 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3543528698/tls.crt::/tmp/serving-cert-3543528698/tls.key\\\\\\\"\\\\nI0202 15:15:36.416658 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 15:15:36.428087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 15:15:36.428134 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 15:15:36.428243 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 15:15:36.428280 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 15:15:36.439678 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 15:15:36.439710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439717 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439722 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 15:15:36.439726 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 15:15:36.439729 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 15:15:36.439732 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 15:15:36.439753 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 15:15:36.450716 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://226f5949820d4f498dbf6fa251c6edf1c6356900bd946e6afae2c29042b1c835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:42Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.434831 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da08f9aecc070b31df2ddc1be7c6eb2ad8ef68a931e57ca2e5c5302f4e75751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:42Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.461188 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" event={"ID":"f5a81ada-e393-499a-aac6-7b40349958bb","Type":"ContainerStarted","Data":"6efaf2216e4c1750139f080a93b4b608737dab2883de5fddfd9a99248d517b2c"} Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.464428 4733 generic.go:334] "Generic (PLEG): container finished" podID="e43cd902-791a-4042-bc08-cf03e075be32" containerID="07e56782051c93b2a320aa539a42764b97647237d228bed2b6af48cee987c60b" exitCode=0 Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.464472 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-nr6ps" event={"ID":"e43cd902-791a-4042-bc08-cf03e075be32","Type":"ContainerDied","Data":"07e56782051c93b2a320aa539a42764b97647237d228bed2b6af48cee987c60b"} Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.467811 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.467840 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.467850 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.467864 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.467875 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:42Z","lastTransitionTime":"2026-02-02T15:15:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.473206 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8c597" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76cc68c3-9e90-45de-87b2-6ea1b68eadce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ac57831c1c880e64aaa7bedcf999133ee09fc5ecc5124b9d54ecedcc9a13c28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwr7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8c597\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:42Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.527700 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8206eca4-7d8b-4b59-a4c1-24631dd8cd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41d4a7bdc21f00b5039385d7227d8cb0f4ea3051d3b5065c71a49cb0bfad26a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c11b1859bf3bc0ec4d5e6a58d9becc1ac6333a1ed9820d83d634f1e93221c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c08b087d85be9bb83007c7e38a5de9182a5091716fb8a2ca5ded1b8191bebc40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fd4969f0f16c9aa44208d2bf578f77d9d50914a8cb1764507175aa2eb58e733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59175c8e5a7f3536168a69fd944bf3d536bcd7281eee96002343b7836a54401b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:42Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.559204 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nr6ps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e43cd902-791a-4042-bc08-cf03e075be32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nr6ps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:42Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.569563 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.569601 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.569612 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.569626 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.569636 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:42Z","lastTransitionTime":"2026-02-02T15:15:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.602935 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5a81ada-e393-499a-aac6-7b40349958bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxvkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:42Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.634221 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:42Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.670093 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca57dabf1d1ee5c6fc29291b1be687ca7e97c72650084b02067057fccac41a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:42Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.671453 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.671495 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.671509 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.671529 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.671540 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:42Z","lastTransitionTime":"2026-02-02T15:15:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.731293 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:42Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.751383 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5xgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56681968-f264-440e-aa12-d4f81879efe9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e250891b1065a38a448aa6abd300b37543e819e87f006b03cc37983cb643982a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdjjl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5xgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:42Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.773602 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.773652 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.773661 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.773680 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.773689 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:42Z","lastTransitionTime":"2026-02-02T15:15:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.787927 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7r9fd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4dfd44b-b363-4162-9f43-9a8dc4a3b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7r9fd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:42Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.831832 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7082da2-f2e7-4b73-ba35-504e5cd4c920\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd383d69b9b27d52fa8672ff6e0294a74876d676973bdc0ca95125fef5f2d085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037e379a109ce230098e3ca223fd1dea3149de3ebe601f56e2bb295f3093a3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6fdd19dac506f2d5d7eedc44ec2cda92a9381b73e91a45d949e4fcde4dfc283\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13965f5abcfb0a36c33310518955f192673e630a1aaba94b49400026de5bc2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://992f4198192cbe4b1c89d575aa4a20c29c1115fd5eaa8986c6ed70c1d6b143c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0202 15:15:31.010033 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 15:15:31.011243 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3543528698/tls.crt::/tmp/serving-cert-3543528698/tls.key\\\\\\\"\\\\nI0202 15:15:36.416658 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 15:15:36.428087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 15:15:36.428134 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 15:15:36.428243 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 15:15:36.428280 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 15:15:36.439678 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 15:15:36.439710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439717 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439722 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 15:15:36.439726 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 15:15:36.439729 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 15:15:36.439732 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 15:15:36.439753 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 15:15:36.450716 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://226f5949820d4f498dbf6fa251c6edf1c6356900bd946e6afae2c29042b1c835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:42Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.872235 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da08f9aecc070b31df2ddc1be7c6eb2ad8ef68a931e57ca2e5c5302f4e75751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:42Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.875919 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.875957 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.875969 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.875985 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.875996 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:42Z","lastTransitionTime":"2026-02-02T15:15:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.911733 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8c597" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76cc68c3-9e90-45de-87b2-6ea1b68eadce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ac57831c1c880e64aaa7bedcf999133ee09fc5ecc5124b9d54ecedcc9a13c28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwr7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8c597\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:42Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.964650 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8206eca4-7d8b-4b59-a4c1-24631dd8cd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41d4a7bdc21f00b5039385d7227d8cb0f4ea3051d3b5065c71a49cb0bfad26a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c11b1859bf3bc0ec4d5e6a58d9becc1ac6333a1ed9820d83d634f1e93221c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c08b087d85be9bb83007c7e38a5de9182a5091716fb8a2ca5ded1b8191bebc40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fd4969f0f16c9aa44208d2bf578f77d9d50914a8cb1764507175aa2eb58e733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59175c8e5a7f3536168a69fd944bf3d536bcd7281eee96002343b7836a54401b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:42Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.978715 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.978758 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.978766 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.978781 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.978790 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:42Z","lastTransitionTime":"2026-02-02T15:15:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:42 crc kubenswrapper[4733]: I0202 15:15:42.993565 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nr6ps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e43cd902-791a-4042-bc08-cf03e075be32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07e56782051c93b2a320aa539a42764b97647237d228bed2b6af48cee987c60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07e56782051c93b2a320aa539a42764b97647237d228bed2b6af48cee987c60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nr6ps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:42Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.046763 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5a81ada-e393-499a-aac6-7b40349958bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxvkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:43Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.075008 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"402a6dc5-6bc7-4d7c-a909-9bb4ca48365d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d5b032497b592363a03f89cb31ba1b551f479004072dddffce540c41190ab81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a3fd02d55a10647428a05a9f53a1429a9827059c523d5deab3c4581c8f1b933\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11600ada13fbc7db1831de385202bc1f041efb656cc58aa49b0f71fe902db9bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fd39f5a05dec7a756dfd519f5df3ef7c77bbb5b8633c5af18a09ff8029cb079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:43Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.081809 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.081872 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.081890 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.081914 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.081931 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:43Z","lastTransitionTime":"2026-02-02T15:15:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.111915 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83d0e810-ea5c-47aa-947b-7865b045bf94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6144eb32f88bc106f7724e48133b568a137c4f3fbbfe30c7fcef405d3ab735a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51318ad71b42f4cc8d376aa50e44580e68da89d83d80defb996329a2ee5c2ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zgsv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:43Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.156417 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:43Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.185562 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.185635 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.185652 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.185679 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.185698 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:43Z","lastTransitionTime":"2026-02-02T15:15:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.200853 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b38b8410bea95465ee58803c4160039bf7c082ca47893680da7469418f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27003b336112760daa2079f4819641850c27fbc8415b63b5094c86195291aca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:43Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.206497 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-07 15:31:38.789725595 +0000 UTC Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.237733 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jh6vk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76a05d6-d9c7-4183-ad90-b3131f225841\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f2659e2a914d3191daddbfb5ad109e487f03dabb559d324244e127add7d41d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jh6vk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:43Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.254934 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.255015 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.254948 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:15:43 crc kubenswrapper[4733]: E0202 15:15:43.255285 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.255298 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:15:43 crc kubenswrapper[4733]: E0202 15:15:43.255669 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:15:43 crc kubenswrapper[4733]: E0202 15:15:43.255502 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:15:43 crc kubenswrapper[4733]: E0202 15:15:43.255880 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.289426 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.289579 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.289602 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.289632 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.289655 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:43Z","lastTransitionTime":"2026-02-02T15:15:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.393015 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.393080 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.393102 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.393141 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.393159 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:43Z","lastTransitionTime":"2026-02-02T15:15:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.476900 4733 generic.go:334] "Generic (PLEG): container finished" podID="e43cd902-791a-4042-bc08-cf03e075be32" containerID="ce850ac7bd4c2444b65d87232114d55b78d11c6f973537c615900092a45d55e8" exitCode=0 Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.476982 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-nr6ps" event={"ID":"e43cd902-791a-4042-bc08-cf03e075be32","Type":"ContainerDied","Data":"ce850ac7bd4c2444b65d87232114d55b78d11c6f973537c615900092a45d55e8"} Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.495940 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.496122 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.496523 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.496858 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.497156 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:43Z","lastTransitionTime":"2026-02-02T15:15:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.502500 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"402a6dc5-6bc7-4d7c-a909-9bb4ca48365d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d5b032497b592363a03f89cb31ba1b551f479004072dddffce540c41190ab81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a3fd02d55a10647428a05a9f53a1429a9827059c523d5deab3c4581c8f1b933\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11600ada13fbc7db1831de385202bc1f041efb656cc58aa49b0f71fe902db9bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fd39f5a05dec7a756dfd519f5df3ef7c77bbb5b8633c5af18a09ff8029cb079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:43Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.524409 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b38b8410bea95465ee58803c4160039bf7c082ca47893680da7469418f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27003b336112760daa2079f4819641850c27fbc8415b63b5094c86195291aca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:43Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.547800 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jh6vk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76a05d6-d9c7-4183-ad90-b3131f225841\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f2659e2a914d3191daddbfb5ad109e487f03dabb559d324244e127add7d41d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jh6vk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:43Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.564665 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83d0e810-ea5c-47aa-947b-7865b045bf94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6144eb32f88bc106f7724e48133b568a137c4f3fbbfe30c7fcef405d3ab735a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51318ad71b42f4cc8d376aa50e44580e68da89d83d80defb996329a2ee5c2ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zgsv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:43Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.582053 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:43Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.599928 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da08f9aecc070b31df2ddc1be7c6eb2ad8ef68a931e57ca2e5c5302f4e75751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:43Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.601976 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.602012 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.602022 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.602037 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.602047 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:43Z","lastTransitionTime":"2026-02-02T15:15:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.614664 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:43Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.625069 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca57dabf1d1ee5c6fc29291b1be687ca7e97c72650084b02067057fccac41a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:43Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.637893 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:43Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.650249 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5xgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56681968-f264-440e-aa12-d4f81879efe9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e250891b1065a38a448aa6abd300b37543e819e87f006b03cc37983cb643982a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdjjl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5xgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:43Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.668321 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7r9fd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4dfd44b-b363-4162-9f43-9a8dc4a3b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7r9fd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:43Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.705263 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.705303 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.705314 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.705333 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.705345 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:43Z","lastTransitionTime":"2026-02-02T15:15:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.713632 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7082da2-f2e7-4b73-ba35-504e5cd4c920\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd383d69b9b27d52fa8672ff6e0294a74876d676973bdc0ca95125fef5f2d085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037e379a109ce230098e3ca223fd1dea3149de3ebe601f56e2bb295f3093a3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6fdd19dac506f2d5d7eedc44ec2cda92a9381b73e91a45d949e4fcde4dfc283\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13965f5abcfb0a36c33310518955f192673e630a1aaba94b49400026de5bc2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://992f4198192cbe4b1c89d575aa4a20c29c1115fd5eaa8986c6ed70c1d6b143c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0202 15:15:31.010033 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 15:15:31.011243 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3543528698/tls.crt::/tmp/serving-cert-3543528698/tls.key\\\\\\\"\\\\nI0202 15:15:36.416658 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 15:15:36.428087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 15:15:36.428134 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 15:15:36.428243 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 15:15:36.428280 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 15:15:36.439678 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 15:15:36.439710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439717 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439722 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 15:15:36.439726 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 15:15:36.439729 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 15:15:36.439732 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 15:15:36.439753 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 15:15:36.450716 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://226f5949820d4f498dbf6fa251c6edf1c6356900bd946e6afae2c29042b1c835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:43Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.753711 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nr6ps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e43cd902-791a-4042-bc08-cf03e075be32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07e56782051c93b2a320aa539a42764b97647237d228bed2b6af48cee987c60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07e56782051c93b2a320aa539a42764b97647237d228bed2b6af48cee987c60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce850ac7bd4c2444b65d87232114d55b78d11c6f973537c615900092a45d55e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce850ac7bd4c2444b65d87232114d55b78d11c6f973537c615900092a45d55e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nr6ps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:43Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.803543 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5a81ada-e393-499a-aac6-7b40349958bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxvkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:43Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.808499 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.808555 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.808572 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.808599 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.808622 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:43Z","lastTransitionTime":"2026-02-02T15:15:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.831474 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8c597" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76cc68c3-9e90-45de-87b2-6ea1b68eadce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ac57831c1c880e64aaa7bedcf999133ee09fc5ecc5124b9d54ecedcc9a13c28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwr7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8c597\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:43Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.877924 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8206eca4-7d8b-4b59-a4c1-24631dd8cd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41d4a7bdc21f00b5039385d7227d8cb0f4ea3051d3b5065c71a49cb0bfad26a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c11b1859bf3bc0ec4d5e6a58d9becc1ac6333a1ed9820d83d634f1e93221c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c08b087d85be9bb83007c7e38a5de9182a5091716fb8a2ca5ded1b8191bebc40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fd4969f0f16c9aa44208d2bf578f77d9d50914a8cb1764507175aa2eb58e733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59175c8e5a7f3536168a69fd944bf3d536bcd7281eee96002343b7836a54401b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:43Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.911216 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.911293 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.911318 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.911349 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:43 crc kubenswrapper[4733]: I0202 15:15:43.911372 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:43Z","lastTransitionTime":"2026-02-02T15:15:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.014021 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.014392 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.014400 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.014413 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.014421 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:44Z","lastTransitionTime":"2026-02-02T15:15:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.117485 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.117543 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.117561 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.117585 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.117603 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:44Z","lastTransitionTime":"2026-02-02T15:15:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.206963 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-11 15:15:36.39797014 +0000 UTC Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.220728 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.220772 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.220781 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.220799 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.220809 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:44Z","lastTransitionTime":"2026-02-02T15:15:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.323961 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.324001 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.324010 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.324028 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.324040 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:44Z","lastTransitionTime":"2026-02-02T15:15:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.427002 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.427062 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.427077 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.427109 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.427128 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:44Z","lastTransitionTime":"2026-02-02T15:15:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.487429 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-nr6ps" event={"ID":"e43cd902-791a-4042-bc08-cf03e075be32","Type":"ContainerStarted","Data":"c2a89a5d1fc43696f33d313aee5dca316a249fdef6a23ff655e31f14cc449cbd"} Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.496113 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" event={"ID":"f5a81ada-e393-499a-aac6-7b40349958bb","Type":"ContainerStarted","Data":"dd866027caf308b856a71578e47f085bdfb2acef5983d1af659102fb202c6bb0"} Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.496563 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.496625 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.520680 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8206eca4-7d8b-4b59-a4c1-24631dd8cd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41d4a7bdc21f00b5039385d7227d8cb0f4ea3051d3b5065c71a49cb0bfad26a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c11b1859bf3bc0ec4d5e6a58d9becc1ac6333a1ed9820d83d634f1e93221c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c08b087d85be9bb83007c7e38a5de9182a5091716fb8a2ca5ded1b8191bebc40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fd4969f0f16c9aa44208d2bf578f77d9d50914a8cb1764507175aa2eb58e733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59175c8e5a7f3536168a69fd944bf3d536bcd7281eee96002343b7836a54401b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:44Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.530426 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.530467 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.530481 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.530505 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.530519 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:44Z","lastTransitionTime":"2026-02-02T15:15:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.540968 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nr6ps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e43cd902-791a-4042-bc08-cf03e075be32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2a89a5d1fc43696f33d313aee5dca316a249fdef6a23ff655e31f14cc449cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07e56782051c93b2a320aa539a42764b97647237d228bed2b6af48cee987c60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07e56782051c93b2a320aa539a42764b97647237d228bed2b6af48cee987c60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce850ac7bd4c2444b65d87232114d55b78d11c6f973537c615900092a45d55e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce850ac7bd4c2444b65d87232114d55b78d11c6f973537c615900092a45d55e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nr6ps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:44Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.543986 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.544836 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.565512 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5a81ada-e393-499a-aac6-7b40349958bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxvkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:44Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.582618 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8c597" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76cc68c3-9e90-45de-87b2-6ea1b68eadce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ac57831c1c880e64aaa7bedcf999133ee09fc5ecc5124b9d54ecedcc9a13c28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwr7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8c597\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:44Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.602439 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"402a6dc5-6bc7-4d7c-a909-9bb4ca48365d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d5b032497b592363a03f89cb31ba1b551f479004072dddffce540c41190ab81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a3fd02d55a10647428a05a9f53a1429a9827059c523d5deab3c4581c8f1b933\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11600ada13fbc7db1831de385202bc1f041efb656cc58aa49b0f71fe902db9bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fd39f5a05dec7a756dfd519f5df3ef7c77bbb5b8633c5af18a09ff8029cb079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:44Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.620048 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:44Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.634518 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.634587 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.634605 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.634630 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.634649 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:44Z","lastTransitionTime":"2026-02-02T15:15:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.644784 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b38b8410bea95465ee58803c4160039bf7c082ca47893680da7469418f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27003b336112760daa2079f4819641850c27fbc8415b63b5094c86195291aca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:44Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.661690 4733 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.666048 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jh6vk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76a05d6-d9c7-4183-ad90-b3131f225841\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f2659e2a914d3191daddbfb5ad109e487f03dabb559d324244e127add7d41d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jh6vk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:44Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.686590 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83d0e810-ea5c-47aa-947b-7865b045bf94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6144eb32f88bc106f7724e48133b568a137c4f3fbbfe30c7fcef405d3ab735a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51318ad71b42f4cc8d376aa50e44580e68da89d83d80defb996329a2ee5c2ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zgsv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:44Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.712144 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca57dabf1d1ee5c6fc29291b1be687ca7e97c72650084b02067057fccac41a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:44Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.737730 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.737783 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.737800 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.737823 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.737841 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:44Z","lastTransitionTime":"2026-02-02T15:15:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.741513 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:44Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.758469 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5xgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56681968-f264-440e-aa12-d4f81879efe9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e250891b1065a38a448aa6abd300b37543e819e87f006b03cc37983cb643982a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdjjl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5xgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:44Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.773344 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7r9fd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4dfd44b-b363-4162-9f43-9a8dc4a3b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7r9fd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:44Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.790540 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7082da2-f2e7-4b73-ba35-504e5cd4c920\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd383d69b9b27d52fa8672ff6e0294a74876d676973bdc0ca95125fef5f2d085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037e379a109ce230098e3ca223fd1dea3149de3ebe601f56e2bb295f3093a3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6fdd19dac506f2d5d7eedc44ec2cda92a9381b73e91a45d949e4fcde4dfc283\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13965f5abcfb0a36c33310518955f192673e630a1aaba94b49400026de5bc2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://992f4198192cbe4b1c89d575aa4a20c29c1115fd5eaa8986c6ed70c1d6b143c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0202 15:15:31.010033 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 15:15:31.011243 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3543528698/tls.crt::/tmp/serving-cert-3543528698/tls.key\\\\\\\"\\\\nI0202 15:15:36.416658 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 15:15:36.428087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 15:15:36.428134 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 15:15:36.428243 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 15:15:36.428280 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 15:15:36.439678 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 15:15:36.439710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439717 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439722 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 15:15:36.439726 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 15:15:36.439729 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 15:15:36.439732 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 15:15:36.439753 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 15:15:36.450716 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://226f5949820d4f498dbf6fa251c6edf1c6356900bd946e6afae2c29042b1c835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:44Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.810703 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da08f9aecc070b31df2ddc1be7c6eb2ad8ef68a931e57ca2e5c5302f4e75751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:44Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.827875 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:44Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.840387 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.840437 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.840449 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.840470 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.840484 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:44Z","lastTransitionTime":"2026-02-02T15:15:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.843654 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"402a6dc5-6bc7-4d7c-a909-9bb4ca48365d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d5b032497b592363a03f89cb31ba1b551f479004072dddffce540c41190ab81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a3fd02d55a10647428a05a9f53a1429a9827059c523d5deab3c4581c8f1b933\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11600ada13fbc7db1831de385202bc1f041efb656cc58aa49b0f71fe902db9bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fd39f5a05dec7a756dfd519f5df3ef7c77bbb5b8633c5af18a09ff8029cb079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:44Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.845888 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.846070 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:15:44 crc kubenswrapper[4733]: E0202 15:15:44.846110 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 15:15:52.846085752 +0000 UTC m=+36.297547120 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:15:44 crc kubenswrapper[4733]: E0202 15:15:44.846223 4733 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 02 15:15:44 crc kubenswrapper[4733]: E0202 15:15:44.846279 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-02 15:15:52.846263787 +0000 UTC m=+36.297725165 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.859888 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:44Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.874553 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b38b8410bea95465ee58803c4160039bf7c082ca47893680da7469418f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27003b336112760daa2079f4819641850c27fbc8415b63b5094c86195291aca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:44Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.889398 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jh6vk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76a05d6-d9c7-4183-ad90-b3131f225841\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f2659e2a914d3191daddbfb5ad109e487f03dabb559d324244e127add7d41d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jh6vk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:44Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.903670 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83d0e810-ea5c-47aa-947b-7865b045bf94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6144eb32f88bc106f7724e48133b568a137c4f3fbbfe30c7fcef405d3ab735a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51318ad71b42f4cc8d376aa50e44580e68da89d83d80defb996329a2ee5c2ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zgsv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:44Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.921487 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7082da2-f2e7-4b73-ba35-504e5cd4c920\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd383d69b9b27d52fa8672ff6e0294a74876d676973bdc0ca95125fef5f2d085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037e379a109ce230098e3ca223fd1dea3149de3ebe601f56e2bb295f3093a3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6fdd19dac506f2d5d7eedc44ec2cda92a9381b73e91a45d949e4fcde4dfc283\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13965f5abcfb0a36c33310518955f192673e630a1aaba94b49400026de5bc2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://992f4198192cbe4b1c89d575aa4a20c29c1115fd5eaa8986c6ed70c1d6b143c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0202 15:15:31.010033 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 15:15:31.011243 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3543528698/tls.crt::/tmp/serving-cert-3543528698/tls.key\\\\\\\"\\\\nI0202 15:15:36.416658 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 15:15:36.428087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 15:15:36.428134 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 15:15:36.428243 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 15:15:36.428280 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 15:15:36.439678 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 15:15:36.439710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439717 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439722 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 15:15:36.439726 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 15:15:36.439729 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 15:15:36.439732 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 15:15:36.439753 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 15:15:36.450716 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://226f5949820d4f498dbf6fa251c6edf1c6356900bd946e6afae2c29042b1c835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:44Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.937062 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da08f9aecc070b31df2ddc1be7c6eb2ad8ef68a931e57ca2e5c5302f4e75751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:44Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.942873 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.942913 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.942929 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.942950 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.942963 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:44Z","lastTransitionTime":"2026-02-02T15:15:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.947527 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d4dfd44b-b363-4162-9f43-9a8dc4a3b110-metrics-certs\") pod \"network-metrics-daemon-7r9fd\" (UID: \"d4dfd44b-b363-4162-9f43-9a8dc4a3b110\") " pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.947581 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.947633 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.947844 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:15:44 crc kubenswrapper[4733]: E0202 15:15:44.947753 4733 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 02 15:15:44 crc kubenswrapper[4733]: E0202 15:15:44.947926 4733 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 02 15:15:44 crc kubenswrapper[4733]: E0202 15:15:44.947783 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 02 15:15:44 crc kubenswrapper[4733]: E0202 15:15:44.947996 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d4dfd44b-b363-4162-9f43-9a8dc4a3b110-metrics-certs podName:d4dfd44b-b363-4162-9f43-9a8dc4a3b110 nodeName:}" failed. No retries permitted until 2026-02-02 15:15:52.947968636 +0000 UTC m=+36.399430004 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d4dfd44b-b363-4162-9f43-9a8dc4a3b110-metrics-certs") pod "network-metrics-daemon-7r9fd" (UID: "d4dfd44b-b363-4162-9f43-9a8dc4a3b110") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 02 15:15:44 crc kubenswrapper[4733]: E0202 15:15:44.948019 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-02 15:15:52.948009767 +0000 UTC m=+36.399471135 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 02 15:15:44 crc kubenswrapper[4733]: E0202 15:15:44.948353 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 02 15:15:44 crc kubenswrapper[4733]: E0202 15:15:44.948376 4733 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 15:15:44 crc kubenswrapper[4733]: E0202 15:15:44.948456 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-02 15:15:52.948441788 +0000 UTC m=+36.399903156 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 15:15:44 crc kubenswrapper[4733]: E0202 15:15:44.947785 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 02 15:15:44 crc kubenswrapper[4733]: E0202 15:15:44.948484 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 02 15:15:44 crc kubenswrapper[4733]: E0202 15:15:44.948495 4733 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 15:15:44 crc kubenswrapper[4733]: E0202 15:15:44.948530 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-02 15:15:52.94852174 +0000 UTC m=+36.399983108 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.956414 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:44Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.974470 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca57dabf1d1ee5c6fc29291b1be687ca7e97c72650084b02067057fccac41a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:44Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:44 crc kubenswrapper[4733]: I0202 15:15:44.992611 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:44Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.011296 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5xgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56681968-f264-440e-aa12-d4f81879efe9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e250891b1065a38a448aa6abd300b37543e819e87f006b03cc37983cb643982a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdjjl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5xgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:45Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.023945 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7r9fd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4dfd44b-b363-4162-9f43-9a8dc4a3b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7r9fd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:45Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.046494 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.046547 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.046559 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.046579 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.046593 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:45Z","lastTransitionTime":"2026-02-02T15:15:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.068716 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8206eca4-7d8b-4b59-a4c1-24631dd8cd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41d4a7bdc21f00b5039385d7227d8cb0f4ea3051d3b5065c71a49cb0bfad26a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c11b1859bf3bc0ec4d5e6a58d9becc1ac6333a1ed9820d83d634f1e93221c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c08b087d85be9bb83007c7e38a5de9182a5091716fb8a2ca5ded1b8191bebc40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fd4969f0f16c9aa44208d2bf578f77d9d50914a8cb1764507175aa2eb58e733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59175c8e5a7f3536168a69fd944bf3d536bcd7281eee96002343b7836a54401b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:45Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.100129 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nr6ps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e43cd902-791a-4042-bc08-cf03e075be32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2a89a5d1fc43696f33d313aee5dca316a249fdef6a23ff655e31f14cc449cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07e56782051c93b2a320aa539a42764b97647237d228bed2b6af48cee987c60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07e56782051c93b2a320aa539a42764b97647237d228bed2b6af48cee987c60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce850ac7bd4c2444b65d87232114d55b78d11c6f973537c615900092a45d55e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce850ac7bd4c2444b65d87232114d55b78d11c6f973537c615900092a45d55e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nr6ps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:45Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.148016 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5a81ada-e393-499a-aac6-7b40349958bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4acf2398e2e19bd54f482cb2273c54738e61627187860bb0b6daadddc2565b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831d85055203f153381eb51d98af0cd2e4c4a0309f837545ab3d855a680133e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1380122d1694c1286f1102e6396c6504b91f63c6eb157689dfc0ea7188c17f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5835cb53385359e9c58eaa5d4da2569c313e0fc265ad79a6c124e8cecb1a734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd6dc4bf5b9fd31bfbee6d9340fe981c0c5fca03a79cef9316340ff5a8b3d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db1db75e392ebea8b387b85cabc4276d553bf2c06694cf7f294a383909f32826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd866027caf308b856a71578e47f085bdfb2acef5983d1af659102fb202c6bb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efaf2216e4c1750139f080a93b4b608737dab2883de5fddfd9a99248d517b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxvkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:45Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.149344 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.149403 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.149421 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.149449 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.149468 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:45Z","lastTransitionTime":"2026-02-02T15:15:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.172577 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8c597" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76cc68c3-9e90-45de-87b2-6ea1b68eadce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ac57831c1c880e64aaa7bedcf999133ee09fc5ecc5124b9d54ecedcc9a13c28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwr7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8c597\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:45Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.207804 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-14 13:39:34.212500598 +0000 UTC Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.252226 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.252286 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.252304 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.252331 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.252348 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:45Z","lastTransitionTime":"2026-02-02T15:15:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.254548 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.254592 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.254592 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:15:45 crc kubenswrapper[4733]: E0202 15:15:45.254729 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.254766 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:15:45 crc kubenswrapper[4733]: E0202 15:15:45.254904 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:15:45 crc kubenswrapper[4733]: E0202 15:15:45.255003 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:15:45 crc kubenswrapper[4733]: E0202 15:15:45.255182 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.355025 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.355080 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.355099 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.355125 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.355144 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:45Z","lastTransitionTime":"2026-02-02T15:15:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.457941 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.458032 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.458048 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.458073 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.458090 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:45Z","lastTransitionTime":"2026-02-02T15:15:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.500725 4733 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.560957 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.561015 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.561033 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.561056 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.561074 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:45Z","lastTransitionTime":"2026-02-02T15:15:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.664641 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.664697 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.664715 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.664740 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.664757 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:45Z","lastTransitionTime":"2026-02-02T15:15:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.767578 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.767638 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.767656 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.767680 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.767698 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:45Z","lastTransitionTime":"2026-02-02T15:15:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.871128 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.871230 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.871251 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.871286 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.871306 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:45Z","lastTransitionTime":"2026-02-02T15:15:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.979761 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.979810 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.979827 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.979850 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.979867 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:45Z","lastTransitionTime":"2026-02-02T15:15:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:45 crc kubenswrapper[4733]: I0202 15:15:45.998013 4733 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.104386 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.104483 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.104500 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.104523 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.104542 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:46Z","lastTransitionTime":"2026-02-02T15:15:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.188751 4733 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.207310 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.207366 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.207383 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.207404 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.207421 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:46Z","lastTransitionTime":"2026-02-02T15:15:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.208677 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-08 12:35:44.778887438 +0000 UTC Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.310381 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.310418 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.310430 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.310470 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.310482 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:46Z","lastTransitionTime":"2026-02-02T15:15:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.413444 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.413519 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.413542 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.413576 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.413600 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:46Z","lastTransitionTime":"2026-02-02T15:15:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.503575 4733 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.516314 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.516358 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.516375 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.516393 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.516405 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:46Z","lastTransitionTime":"2026-02-02T15:15:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.618938 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.618989 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.619003 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.619023 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.619037 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:46Z","lastTransitionTime":"2026-02-02T15:15:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.722236 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.722298 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.722318 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.722343 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.722361 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:46Z","lastTransitionTime":"2026-02-02T15:15:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.723913 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.724048 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.724133 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.724248 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.724326 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:46Z","lastTransitionTime":"2026-02-02T15:15:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:46 crc kubenswrapper[4733]: E0202 15:15:46.738110 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd742269-802e-45ec-a88e-17aee17c574e\\\",\\\"systemUUID\\\":\\\"91b64b75-3cd9-4fb9-93f0-dbd3bcdcaec4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:46Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.741504 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.741566 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.741581 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.741598 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.741611 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:46Z","lastTransitionTime":"2026-02-02T15:15:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:46 crc kubenswrapper[4733]: E0202 15:15:46.752662 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd742269-802e-45ec-a88e-17aee17c574e\\\",\\\"systemUUID\\\":\\\"91b64b75-3cd9-4fb9-93f0-dbd3bcdcaec4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:46Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.756611 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.756674 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.756690 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.756707 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.756719 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:46Z","lastTransitionTime":"2026-02-02T15:15:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:46 crc kubenswrapper[4733]: E0202 15:15:46.773412 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd742269-802e-45ec-a88e-17aee17c574e\\\",\\\"systemUUID\\\":\\\"91b64b75-3cd9-4fb9-93f0-dbd3bcdcaec4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:46Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.777204 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.777298 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.777315 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.777337 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.777504 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:46Z","lastTransitionTime":"2026-02-02T15:15:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:46 crc kubenswrapper[4733]: E0202 15:15:46.792736 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd742269-802e-45ec-a88e-17aee17c574e\\\",\\\"systemUUID\\\":\\\"91b64b75-3cd9-4fb9-93f0-dbd3bcdcaec4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:46Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.796728 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.796790 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.796806 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.796828 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.796843 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:46Z","lastTransitionTime":"2026-02-02T15:15:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:46 crc kubenswrapper[4733]: E0202 15:15:46.814953 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd742269-802e-45ec-a88e-17aee17c574e\\\",\\\"systemUUID\\\":\\\"91b64b75-3cd9-4fb9-93f0-dbd3bcdcaec4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:46Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:46 crc kubenswrapper[4733]: E0202 15:15:46.815110 4733 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.824406 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.824471 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.824483 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.824503 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.824516 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:46Z","lastTransitionTime":"2026-02-02T15:15:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.927352 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.927398 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.927417 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.927439 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:46 crc kubenswrapper[4733]: I0202 15:15:46.927455 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:46Z","lastTransitionTime":"2026-02-02T15:15:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.030335 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.030386 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.030398 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.030420 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.030433 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:47Z","lastTransitionTime":"2026-02-02T15:15:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.132450 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.132491 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.132500 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.132514 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.132525 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:47Z","lastTransitionTime":"2026-02-02T15:15:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.209313 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-03 14:26:31.08874114 +0000 UTC Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.235565 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.235608 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.235619 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.235635 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.235647 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:47Z","lastTransitionTime":"2026-02-02T15:15:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.255334 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.255362 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.255395 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:15:47 crc kubenswrapper[4733]: E0202 15:15:47.255500 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:15:47 crc kubenswrapper[4733]: E0202 15:15:47.255599 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:15:47 crc kubenswrapper[4733]: E0202 15:15:47.255699 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.255958 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:15:47 crc kubenswrapper[4733]: E0202 15:15:47.256069 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.268900 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"402a6dc5-6bc7-4d7c-a909-9bb4ca48365d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d5b032497b592363a03f89cb31ba1b551f479004072dddffce540c41190ab81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a3fd02d55a10647428a05a9f53a1429a9827059c523d5deab3c4581c8f1b933\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11600ada13fbc7db1831de385202bc1f041efb656cc58aa49b0f71fe902db9bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fd39f5a05dec7a756dfd519f5df3ef7c77bbb5b8633c5af18a09ff8029cb079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:47Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.284576 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:47Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.301720 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b38b8410bea95465ee58803c4160039bf7c082ca47893680da7469418f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27003b336112760daa2079f4819641850c27fbc8415b63b5094c86195291aca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:47Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.314498 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jh6vk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76a05d6-d9c7-4183-ad90-b3131f225841\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f2659e2a914d3191daddbfb5ad109e487f03dabb559d324244e127add7d41d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jh6vk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:47Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.329095 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83d0e810-ea5c-47aa-947b-7865b045bf94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6144eb32f88bc106f7724e48133b568a137c4f3fbbfe30c7fcef405d3ab735a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51318ad71b42f4cc8d376aa50e44580e68da89d83d80defb996329a2ee5c2ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zgsv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:47Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.338300 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.338363 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.338389 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.338417 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.338439 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:47Z","lastTransitionTime":"2026-02-02T15:15:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.344370 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7r9fd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4dfd44b-b363-4162-9f43-9a8dc4a3b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7r9fd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:47Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.363062 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7082da2-f2e7-4b73-ba35-504e5cd4c920\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd383d69b9b27d52fa8672ff6e0294a74876d676973bdc0ca95125fef5f2d085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037e379a109ce230098e3ca223fd1dea3149de3ebe601f56e2bb295f3093a3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6fdd19dac506f2d5d7eedc44ec2cda92a9381b73e91a45d949e4fcde4dfc283\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13965f5abcfb0a36c33310518955f192673e630a1aaba94b49400026de5bc2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://992f4198192cbe4b1c89d575aa4a20c29c1115fd5eaa8986c6ed70c1d6b143c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0202 15:15:31.010033 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 15:15:31.011243 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3543528698/tls.crt::/tmp/serving-cert-3543528698/tls.key\\\\\\\"\\\\nI0202 15:15:36.416658 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 15:15:36.428087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 15:15:36.428134 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 15:15:36.428243 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 15:15:36.428280 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 15:15:36.439678 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 15:15:36.439710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439717 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439722 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 15:15:36.439726 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 15:15:36.439729 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 15:15:36.439732 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 15:15:36.439753 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 15:15:36.450716 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://226f5949820d4f498dbf6fa251c6edf1c6356900bd946e6afae2c29042b1c835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:47Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.382893 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da08f9aecc070b31df2ddc1be7c6eb2ad8ef68a931e57ca2e5c5302f4e75751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:47Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.403699 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:47Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.421599 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca57dabf1d1ee5c6fc29291b1be687ca7e97c72650084b02067057fccac41a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:47Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.433445 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:47Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.441489 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.441526 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.441538 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.441554 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.441566 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:47Z","lastTransitionTime":"2026-02-02T15:15:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.445340 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5xgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56681968-f264-440e-aa12-d4f81879efe9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e250891b1065a38a448aa6abd300b37543e819e87f006b03cc37983cb643982a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdjjl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5xgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:47Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.473123 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8206eca4-7d8b-4b59-a4c1-24631dd8cd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41d4a7bdc21f00b5039385d7227d8cb0f4ea3051d3b5065c71a49cb0bfad26a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c11b1859bf3bc0ec4d5e6a58d9becc1ac6333a1ed9820d83d634f1e93221c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c08b087d85be9bb83007c7e38a5de9182a5091716fb8a2ca5ded1b8191bebc40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fd4969f0f16c9aa44208d2bf578f77d9d50914a8cb1764507175aa2eb58e733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59175c8e5a7f3536168a69fd944bf3d536bcd7281eee96002343b7836a54401b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:47Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.491994 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nr6ps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e43cd902-791a-4042-bc08-cf03e075be32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2a89a5d1fc43696f33d313aee5dca316a249fdef6a23ff655e31f14cc449cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07e56782051c93b2a320aa539a42764b97647237d228bed2b6af48cee987c60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07e56782051c93b2a320aa539a42764b97647237d228bed2b6af48cee987c60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce850ac7bd4c2444b65d87232114d55b78d11c6f973537c615900092a45d55e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce850ac7bd4c2444b65d87232114d55b78d11c6f973537c615900092a45d55e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nr6ps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:47Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.509054 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxvkf_f5a81ada-e393-499a-aac6-7b40349958bb/ovnkube-controller/0.log" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.512713 4733 generic.go:334] "Generic (PLEG): container finished" podID="f5a81ada-e393-499a-aac6-7b40349958bb" containerID="dd866027caf308b856a71578e47f085bdfb2acef5983d1af659102fb202c6bb0" exitCode=1 Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.512766 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" event={"ID":"f5a81ada-e393-499a-aac6-7b40349958bb","Type":"ContainerDied","Data":"dd866027caf308b856a71578e47f085bdfb2acef5983d1af659102fb202c6bb0"} Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.513832 4733 scope.go:117] "RemoveContainer" containerID="dd866027caf308b856a71578e47f085bdfb2acef5983d1af659102fb202c6bb0" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.522922 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5a81ada-e393-499a-aac6-7b40349958bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4acf2398e2e19bd54f482cb2273c54738e61627187860bb0b6daadddc2565b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831d85055203f153381eb51d98af0cd2e4c4a0309f837545ab3d855a680133e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1380122d1694c1286f1102e6396c6504b91f63c6eb157689dfc0ea7188c17f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5835cb53385359e9c58eaa5d4da2569c313e0fc265ad79a6c124e8cecb1a734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd6dc4bf5b9fd31bfbee6d9340fe981c0c5fca03a79cef9316340ff5a8b3d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db1db75e392ebea8b387b85cabc4276d553bf2c06694cf7f294a383909f32826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd866027caf308b856a71578e47f085bdfb2acef5983d1af659102fb202c6bb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efaf2216e4c1750139f080a93b4b608737dab2883de5fddfd9a99248d517b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxvkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:47Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.538774 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8c597" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76cc68c3-9e90-45de-87b2-6ea1b68eadce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ac57831c1c880e64aaa7bedcf999133ee09fc5ecc5124b9d54ecedcc9a13c28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwr7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8c597\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:47Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.544956 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.545046 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.545106 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.545134 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.545211 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:47Z","lastTransitionTime":"2026-02-02T15:15:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.558131 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca57dabf1d1ee5c6fc29291b1be687ca7e97c72650084b02067057fccac41a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:47Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.577070 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:47Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.597703 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5xgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56681968-f264-440e-aa12-d4f81879efe9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e250891b1065a38a448aa6abd300b37543e819e87f006b03cc37983cb643982a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdjjl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5xgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:47Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.614246 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7r9fd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4dfd44b-b363-4162-9f43-9a8dc4a3b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7r9fd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:47Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.637258 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7082da2-f2e7-4b73-ba35-504e5cd4c920\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd383d69b9b27d52fa8672ff6e0294a74876d676973bdc0ca95125fef5f2d085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037e379a109ce230098e3ca223fd1dea3149de3ebe601f56e2bb295f3093a3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6fdd19dac506f2d5d7eedc44ec2cda92a9381b73e91a45d949e4fcde4dfc283\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13965f5abcfb0a36c33310518955f192673e630a1aaba94b49400026de5bc2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://992f4198192cbe4b1c89d575aa4a20c29c1115fd5eaa8986c6ed70c1d6b143c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0202 15:15:31.010033 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 15:15:31.011243 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3543528698/tls.crt::/tmp/serving-cert-3543528698/tls.key\\\\\\\"\\\\nI0202 15:15:36.416658 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 15:15:36.428087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 15:15:36.428134 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 15:15:36.428243 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 15:15:36.428280 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 15:15:36.439678 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 15:15:36.439710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439717 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439722 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 15:15:36.439726 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 15:15:36.439729 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 15:15:36.439732 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 15:15:36.439753 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 15:15:36.450716 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://226f5949820d4f498dbf6fa251c6edf1c6356900bd946e6afae2c29042b1c835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:47Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.647917 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.647966 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.647978 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.647997 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.648011 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:47Z","lastTransitionTime":"2026-02-02T15:15:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.661503 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da08f9aecc070b31df2ddc1be7c6eb2ad8ef68a931e57ca2e5c5302f4e75751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:47Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.690456 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:47Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.707356 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.725880 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8206eca4-7d8b-4b59-a4c1-24631dd8cd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41d4a7bdc21f00b5039385d7227d8cb0f4ea3051d3b5065c71a49cb0bfad26a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c11b1859bf3bc0ec4d5e6a58d9becc1ac6333a1ed9820d83d634f1e93221c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c08b087d85be9bb83007c7e38a5de9182a5091716fb8a2ca5ded1b8191bebc40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fd4969f0f16c9aa44208d2bf578f77d9d50914a8cb1764507175aa2eb58e733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59175c8e5a7f3536168a69fd944bf3d536bcd7281eee96002343b7836a54401b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:47Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.751716 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.751760 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.751777 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.751798 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.751814 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:47Z","lastTransitionTime":"2026-02-02T15:15:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.751857 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nr6ps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e43cd902-791a-4042-bc08-cf03e075be32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2a89a5d1fc43696f33d313aee5dca316a249fdef6a23ff655e31f14cc449cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07e56782051c93b2a320aa539a42764b97647237d228bed2b6af48cee987c60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07e56782051c93b2a320aa539a42764b97647237d228bed2b6af48cee987c60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce850ac7bd4c2444b65d87232114d55b78d11c6f973537c615900092a45d55e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce850ac7bd4c2444b65d87232114d55b78d11c6f973537c615900092a45d55e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nr6ps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:47Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.789684 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5a81ada-e393-499a-aac6-7b40349958bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4acf2398e2e19bd54f482cb2273c54738e61627187860bb0b6daadddc2565b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831d85055203f153381eb51d98af0cd2e4c4a0309f837545ab3d855a680133e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1380122d1694c1286f1102e6396c6504b91f63c6eb157689dfc0ea7188c17f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5835cb53385359e9c58eaa5d4da2569c313e0fc265ad79a6c124e8cecb1a734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd6dc4bf5b9fd31bfbee6d9340fe981c0c5fca03a79cef9316340ff5a8b3d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db1db75e392ebea8b387b85cabc4276d553bf2c06694cf7f294a383909f32826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd866027caf308b856a71578e47f085bdfb2acef5983d1af659102fb202c6bb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd866027caf308b856a71578e47f085bdfb2acef5983d1af659102fb202c6bb0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T15:15:47Z\\\",\\\"message\\\":\\\"ub.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0202 15:15:47.138308 6058 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0202 15:15:47.138443 6058 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0202 15:15:47.138624 6058 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0202 15:15:47.138629 6058 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0202 15:15:47.138693 6058 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0202 15:15:47.139237 6058 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0202 15:15:47.139373 6058 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efaf2216e4c1750139f080a93b4b608737dab2883de5fddfd9a99248d517b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxvkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:47Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.811871 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8c597" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76cc68c3-9e90-45de-87b2-6ea1b68eadce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ac57831c1c880e64aaa7bedcf999133ee09fc5ecc5124b9d54ecedcc9a13c28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwr7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8c597\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:47Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.831713 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"402a6dc5-6bc7-4d7c-a909-9bb4ca48365d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d5b032497b592363a03f89cb31ba1b551f479004072dddffce540c41190ab81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a3fd02d55a10647428a05a9f53a1429a9827059c523d5deab3c4581c8f1b933\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11600ada13fbc7db1831de385202bc1f041efb656cc58aa49b0f71fe902db9bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fd39f5a05dec7a756dfd519f5df3ef7c77bbb5b8633c5af18a09ff8029cb079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:47Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.847712 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:47Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.854016 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.854047 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.854056 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.854073 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.854084 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:47Z","lastTransitionTime":"2026-02-02T15:15:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.862798 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b38b8410bea95465ee58803c4160039bf7c082ca47893680da7469418f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27003b336112760daa2079f4819641850c27fbc8415b63b5094c86195291aca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:47Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.879477 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jh6vk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76a05d6-d9c7-4183-ad90-b3131f225841\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f2659e2a914d3191daddbfb5ad109e487f03dabb559d324244e127add7d41d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jh6vk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:47Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.888765 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83d0e810-ea5c-47aa-947b-7865b045bf94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6144eb32f88bc106f7724e48133b568a137c4f3fbbfe30c7fcef405d3ab735a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51318ad71b42f4cc8d376aa50e44580e68da89d83d80defb996329a2ee5c2ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zgsv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:47Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.956772 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.956827 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.956840 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.956858 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:47 crc kubenswrapper[4733]: I0202 15:15:47.956872 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:47Z","lastTransitionTime":"2026-02-02T15:15:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.080481 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.080532 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.080550 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.080571 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.080587 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:48Z","lastTransitionTime":"2026-02-02T15:15:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.086401 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.112854 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"402a6dc5-6bc7-4d7c-a909-9bb4ca48365d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d5b032497b592363a03f89cb31ba1b551f479004072dddffce540c41190ab81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a3fd02d55a10647428a05a9f53a1429a9827059c523d5deab3c4581c8f1b933\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11600ada13fbc7db1831de385202bc1f041efb656cc58aa49b0f71fe902db9bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fd39f5a05dec7a756dfd519f5df3ef7c77bbb5b8633c5af18a09ff8029cb079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:48Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.129535 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:48Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.146213 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b38b8410bea95465ee58803c4160039bf7c082ca47893680da7469418f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27003b336112760daa2079f4819641850c27fbc8415b63b5094c86195291aca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:48Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.161745 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jh6vk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76a05d6-d9c7-4183-ad90-b3131f225841\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f2659e2a914d3191daddbfb5ad109e487f03dabb559d324244e127add7d41d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jh6vk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:48Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.176809 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83d0e810-ea5c-47aa-947b-7865b045bf94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6144eb32f88bc106f7724e48133b568a137c4f3fbbfe30c7fcef405d3ab735a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51318ad71b42f4cc8d376aa50e44580e68da89d83d80defb996329a2ee5c2ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zgsv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:48Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.182576 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.182615 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.182624 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.182638 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.182648 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:48Z","lastTransitionTime":"2026-02-02T15:15:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.193618 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7082da2-f2e7-4b73-ba35-504e5cd4c920\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd383d69b9b27d52fa8672ff6e0294a74876d676973bdc0ca95125fef5f2d085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037e379a109ce230098e3ca223fd1dea3149de3ebe601f56e2bb295f3093a3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6fdd19dac506f2d5d7eedc44ec2cda92a9381b73e91a45d949e4fcde4dfc283\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13965f5abcfb0a36c33310518955f192673e630a1aaba94b49400026de5bc2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://992f4198192cbe4b1c89d575aa4a20c29c1115fd5eaa8986c6ed70c1d6b143c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0202 15:15:31.010033 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 15:15:31.011243 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3543528698/tls.crt::/tmp/serving-cert-3543528698/tls.key\\\\\\\"\\\\nI0202 15:15:36.416658 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 15:15:36.428087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 15:15:36.428134 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 15:15:36.428243 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 15:15:36.428280 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 15:15:36.439678 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 15:15:36.439710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439717 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439722 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 15:15:36.439726 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 15:15:36.439729 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 15:15:36.439732 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 15:15:36.439753 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 15:15:36.450716 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://226f5949820d4f498dbf6fa251c6edf1c6356900bd946e6afae2c29042b1c835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:48Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.209844 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-12 21:35:13.224591762 +0000 UTC Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.224158 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da08f9aecc070b31df2ddc1be7c6eb2ad8ef68a931e57ca2e5c5302f4e75751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:48Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.238407 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:48Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.250492 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca57dabf1d1ee5c6fc29291b1be687ca7e97c72650084b02067057fccac41a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:48Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.264328 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:48Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.273987 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5xgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56681968-f264-440e-aa12-d4f81879efe9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e250891b1065a38a448aa6abd300b37543e819e87f006b03cc37983cb643982a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdjjl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5xgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:48Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.284945 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.285010 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.285022 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.285043 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.285055 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:48Z","lastTransitionTime":"2026-02-02T15:15:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.287044 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7r9fd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4dfd44b-b363-4162-9f43-9a8dc4a3b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7r9fd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:48Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.306641 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8206eca4-7d8b-4b59-a4c1-24631dd8cd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41d4a7bdc21f00b5039385d7227d8cb0f4ea3051d3b5065c71a49cb0bfad26a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c11b1859bf3bc0ec4d5e6a58d9becc1ac6333a1ed9820d83d634f1e93221c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c08b087d85be9bb83007c7e38a5de9182a5091716fb8a2ca5ded1b8191bebc40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fd4969f0f16c9aa44208d2bf578f77d9d50914a8cb1764507175aa2eb58e733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59175c8e5a7f3536168a69fd944bf3d536bcd7281eee96002343b7836a54401b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:48Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.321092 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nr6ps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e43cd902-791a-4042-bc08-cf03e075be32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2a89a5d1fc43696f33d313aee5dca316a249fdef6a23ff655e31f14cc449cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07e56782051c93b2a320aa539a42764b97647237d228bed2b6af48cee987c60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07e56782051c93b2a320aa539a42764b97647237d228bed2b6af48cee987c60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce850ac7bd4c2444b65d87232114d55b78d11c6f973537c615900092a45d55e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce850ac7bd4c2444b65d87232114d55b78d11c6f973537c615900092a45d55e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nr6ps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:48Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.338396 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5a81ada-e393-499a-aac6-7b40349958bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4acf2398e2e19bd54f482cb2273c54738e61627187860bb0b6daadddc2565b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831d85055203f153381eb51d98af0cd2e4c4a0309f837545ab3d855a680133e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1380122d1694c1286f1102e6396c6504b91f63c6eb157689dfc0ea7188c17f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5835cb53385359e9c58eaa5d4da2569c313e0fc265ad79a6c124e8cecb1a734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd6dc4bf5b9fd31bfbee6d9340fe981c0c5fca03a79cef9316340ff5a8b3d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db1db75e392ebea8b387b85cabc4276d553bf2c06694cf7f294a383909f32826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd866027caf308b856a71578e47f085bdfb2acef5983d1af659102fb202c6bb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd866027caf308b856a71578e47f085bdfb2acef5983d1af659102fb202c6bb0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T15:15:47Z\\\",\\\"message\\\":\\\"ub.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0202 15:15:47.138308 6058 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0202 15:15:47.138443 6058 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0202 15:15:47.138624 6058 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0202 15:15:47.138629 6058 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0202 15:15:47.138693 6058 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0202 15:15:47.139237 6058 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0202 15:15:47.139373 6058 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efaf2216e4c1750139f080a93b4b608737dab2883de5fddfd9a99248d517b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxvkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:48Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.349357 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8c597" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76cc68c3-9e90-45de-87b2-6ea1b68eadce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ac57831c1c880e64aaa7bedcf999133ee09fc5ecc5124b9d54ecedcc9a13c28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwr7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8c597\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:48Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.386918 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.387001 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.387027 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.387095 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.387120 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:48Z","lastTransitionTime":"2026-02-02T15:15:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.489770 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.489818 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.489832 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.489850 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.489862 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:48Z","lastTransitionTime":"2026-02-02T15:15:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.518940 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxvkf_f5a81ada-e393-499a-aac6-7b40349958bb/ovnkube-controller/0.log" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.521273 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" event={"ID":"f5a81ada-e393-499a-aac6-7b40349958bb","Type":"ContainerStarted","Data":"58d0beeefc5a719f002da8a57bb8f5209f8e9e527159fcfbf77c28cf1a30fdf6"} Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.522973 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.548048 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"402a6dc5-6bc7-4d7c-a909-9bb4ca48365d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d5b032497b592363a03f89cb31ba1b551f479004072dddffce540c41190ab81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a3fd02d55a10647428a05a9f53a1429a9827059c523d5deab3c4581c8f1b933\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11600ada13fbc7db1831de385202bc1f041efb656cc58aa49b0f71fe902db9bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fd39f5a05dec7a756dfd519f5df3ef7c77bbb5b8633c5af18a09ff8029cb079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:48Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.565952 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:48Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.580917 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b38b8410bea95465ee58803c4160039bf7c082ca47893680da7469418f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27003b336112760daa2079f4819641850c27fbc8415b63b5094c86195291aca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:48Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.592895 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.592942 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.592951 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.592968 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.592977 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:48Z","lastTransitionTime":"2026-02-02T15:15:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.601925 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jh6vk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76a05d6-d9c7-4183-ad90-b3131f225841\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f2659e2a914d3191daddbfb5ad109e487f03dabb559d324244e127add7d41d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jh6vk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:48Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.619005 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83d0e810-ea5c-47aa-947b-7865b045bf94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6144eb32f88bc106f7724e48133b568a137c4f3fbbfe30c7fcef405d3ab735a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51318ad71b42f4cc8d376aa50e44580e68da89d83d80defb996329a2ee5c2ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zgsv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:48Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.637078 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7082da2-f2e7-4b73-ba35-504e5cd4c920\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd383d69b9b27d52fa8672ff6e0294a74876d676973bdc0ca95125fef5f2d085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037e379a109ce230098e3ca223fd1dea3149de3ebe601f56e2bb295f3093a3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6fdd19dac506f2d5d7eedc44ec2cda92a9381b73e91a45d949e4fcde4dfc283\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13965f5abcfb0a36c33310518955f192673e630a1aaba94b49400026de5bc2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://992f4198192cbe4b1c89d575aa4a20c29c1115fd5eaa8986c6ed70c1d6b143c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0202 15:15:31.010033 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 15:15:31.011243 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3543528698/tls.crt::/tmp/serving-cert-3543528698/tls.key\\\\\\\"\\\\nI0202 15:15:36.416658 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 15:15:36.428087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 15:15:36.428134 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 15:15:36.428243 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 15:15:36.428280 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 15:15:36.439678 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 15:15:36.439710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439717 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439722 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 15:15:36.439726 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 15:15:36.439729 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 15:15:36.439732 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 15:15:36.439753 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 15:15:36.450716 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://226f5949820d4f498dbf6fa251c6edf1c6356900bd946e6afae2c29042b1c835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:48Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.653314 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da08f9aecc070b31df2ddc1be7c6eb2ad8ef68a931e57ca2e5c5302f4e75751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:48Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.667419 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:48Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.679389 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca57dabf1d1ee5c6fc29291b1be687ca7e97c72650084b02067057fccac41a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:48Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.690142 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:48Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.695274 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.695313 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.695325 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.695343 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.695354 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:48Z","lastTransitionTime":"2026-02-02T15:15:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.702767 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5xgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56681968-f264-440e-aa12-d4f81879efe9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e250891b1065a38a448aa6abd300b37543e819e87f006b03cc37983cb643982a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdjjl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5xgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:48Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.712578 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7r9fd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4dfd44b-b363-4162-9f43-9a8dc4a3b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7r9fd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:48Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.730922 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8206eca4-7d8b-4b59-a4c1-24631dd8cd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41d4a7bdc21f00b5039385d7227d8cb0f4ea3051d3b5065c71a49cb0bfad26a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c11b1859bf3bc0ec4d5e6a58d9becc1ac6333a1ed9820d83d634f1e93221c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c08b087d85be9bb83007c7e38a5de9182a5091716fb8a2ca5ded1b8191bebc40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fd4969f0f16c9aa44208d2bf578f77d9d50914a8cb1764507175aa2eb58e733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59175c8e5a7f3536168a69fd944bf3d536bcd7281eee96002343b7836a54401b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:48Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.747550 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nr6ps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e43cd902-791a-4042-bc08-cf03e075be32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2a89a5d1fc43696f33d313aee5dca316a249fdef6a23ff655e31f14cc449cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07e56782051c93b2a320aa539a42764b97647237d228bed2b6af48cee987c60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07e56782051c93b2a320aa539a42764b97647237d228bed2b6af48cee987c60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce850ac7bd4c2444b65d87232114d55b78d11c6f973537c615900092a45d55e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce850ac7bd4c2444b65d87232114d55b78d11c6f973537c615900092a45d55e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nr6ps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:48Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.773444 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5a81ada-e393-499a-aac6-7b40349958bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4acf2398e2e19bd54f482cb2273c54738e61627187860bb0b6daadddc2565b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831d85055203f153381eb51d98af0cd2e4c4a0309f837545ab3d855a680133e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1380122d1694c1286f1102e6396c6504b91f63c6eb157689dfc0ea7188c17f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5835cb53385359e9c58eaa5d4da2569c313e0fc265ad79a6c124e8cecb1a734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd6dc4bf5b9fd31bfbee6d9340fe981c0c5fca03a79cef9316340ff5a8b3d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db1db75e392ebea8b387b85cabc4276d553bf2c06694cf7f294a383909f32826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58d0beeefc5a719f002da8a57bb8f5209f8e9e527159fcfbf77c28cf1a30fdf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd866027caf308b856a71578e47f085bdfb2acef5983d1af659102fb202c6bb0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T15:15:47Z\\\",\\\"message\\\":\\\"ub.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0202 15:15:47.138308 6058 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0202 15:15:47.138443 6058 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0202 15:15:47.138624 6058 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0202 15:15:47.138629 6058 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0202 15:15:47.138693 6058 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0202 15:15:47.139237 6058 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0202 15:15:47.139373 6058 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efaf2216e4c1750139f080a93b4b608737dab2883de5fddfd9a99248d517b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxvkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:48Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.785312 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8c597" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76cc68c3-9e90-45de-87b2-6ea1b68eadce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ac57831c1c880e64aaa7bedcf999133ee09fc5ecc5124b9d54ecedcc9a13c28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwr7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8c597\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:48Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.797669 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.797717 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.797736 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.797758 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.797772 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:48Z","lastTransitionTime":"2026-02-02T15:15:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.901195 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.901253 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.901271 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.901295 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:48 crc kubenswrapper[4733]: I0202 15:15:48.901312 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:48Z","lastTransitionTime":"2026-02-02T15:15:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.004704 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.004770 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.004787 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.004811 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.004828 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:49Z","lastTransitionTime":"2026-02-02T15:15:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.112630 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.112894 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.112970 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.113078 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.113304 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:49Z","lastTransitionTime":"2026-02-02T15:15:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.210622 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-29 14:17:53.338101398 +0000 UTC Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.215932 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.216035 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.216093 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.216150 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.216228 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:49Z","lastTransitionTime":"2026-02-02T15:15:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.254837 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.254878 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.254972 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.255066 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:15:49 crc kubenswrapper[4733]: E0202 15:15:49.255102 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:15:49 crc kubenswrapper[4733]: E0202 15:15:49.255227 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:15:49 crc kubenswrapper[4733]: E0202 15:15:49.255352 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:15:49 crc kubenswrapper[4733]: E0202 15:15:49.255618 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.321101 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.321152 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.321191 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.321217 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.321235 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:49Z","lastTransitionTime":"2026-02-02T15:15:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.424492 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.424547 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.424564 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.424588 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.424605 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:49Z","lastTransitionTime":"2026-02-02T15:15:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.526765 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.526820 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.526899 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.526930 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.526949 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:49Z","lastTransitionTime":"2026-02-02T15:15:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.527855 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxvkf_f5a81ada-e393-499a-aac6-7b40349958bb/ovnkube-controller/1.log" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.528783 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxvkf_f5a81ada-e393-499a-aac6-7b40349958bb/ovnkube-controller/0.log" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.532419 4733 generic.go:334] "Generic (PLEG): container finished" podID="f5a81ada-e393-499a-aac6-7b40349958bb" containerID="58d0beeefc5a719f002da8a57bb8f5209f8e9e527159fcfbf77c28cf1a30fdf6" exitCode=1 Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.532466 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" event={"ID":"f5a81ada-e393-499a-aac6-7b40349958bb","Type":"ContainerDied","Data":"58d0beeefc5a719f002da8a57bb8f5209f8e9e527159fcfbf77c28cf1a30fdf6"} Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.532503 4733 scope.go:117] "RemoveContainer" containerID="dd866027caf308b856a71578e47f085bdfb2acef5983d1af659102fb202c6bb0" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.533148 4733 scope.go:117] "RemoveContainer" containerID="58d0beeefc5a719f002da8a57bb8f5209f8e9e527159fcfbf77c28cf1a30fdf6" Feb 02 15:15:49 crc kubenswrapper[4733]: E0202 15:15:49.533358 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qxvkf_openshift-ovn-kubernetes(f5a81ada-e393-499a-aac6-7b40349958bb)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.570584 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8206eca4-7d8b-4b59-a4c1-24631dd8cd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41d4a7bdc21f00b5039385d7227d8cb0f4ea3051d3b5065c71a49cb0bfad26a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c11b1859bf3bc0ec4d5e6a58d9becc1ac6333a1ed9820d83d634f1e93221c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c08b087d85be9bb83007c7e38a5de9182a5091716fb8a2ca5ded1b8191bebc40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fd4969f0f16c9aa44208d2bf578f77d9d50914a8cb1764507175aa2eb58e733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59175c8e5a7f3536168a69fd944bf3d536bcd7281eee96002343b7836a54401b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:49Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.595369 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nr6ps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e43cd902-791a-4042-bc08-cf03e075be32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2a89a5d1fc43696f33d313aee5dca316a249fdef6a23ff655e31f14cc449cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07e56782051c93b2a320aa539a42764b97647237d228bed2b6af48cee987c60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07e56782051c93b2a320aa539a42764b97647237d228bed2b6af48cee987c60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce850ac7bd4c2444b65d87232114d55b78d11c6f973537c615900092a45d55e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce850ac7bd4c2444b65d87232114d55b78d11c6f973537c615900092a45d55e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nr6ps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:49Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.627720 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5a81ada-e393-499a-aac6-7b40349958bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4acf2398e2e19bd54f482cb2273c54738e61627187860bb0b6daadddc2565b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831d85055203f153381eb51d98af0cd2e4c4a0309f837545ab3d855a680133e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1380122d1694c1286f1102e6396c6504b91f63c6eb157689dfc0ea7188c17f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5835cb53385359e9c58eaa5d4da2569c313e0fc265ad79a6c124e8cecb1a734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd6dc4bf5b9fd31bfbee6d9340fe981c0c5fca03a79cef9316340ff5a8b3d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db1db75e392ebea8b387b85cabc4276d553bf2c06694cf7f294a383909f32826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58d0beeefc5a719f002da8a57bb8f5209f8e9e527159fcfbf77c28cf1a30fdf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd866027caf308b856a71578e47f085bdfb2acef5983d1af659102fb202c6bb0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T15:15:47Z\\\",\\\"message\\\":\\\"ub.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0202 15:15:47.138308 6058 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0202 15:15:47.138443 6058 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0202 15:15:47.138624 6058 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0202 15:15:47.138629 6058 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0202 15:15:47.138693 6058 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0202 15:15:47.139237 6058 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0202 15:15:47.139373 6058 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58d0beeefc5a719f002da8a57bb8f5209f8e9e527159fcfbf77c28cf1a30fdf6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T15:15:48Z\\\",\\\"message\\\":\\\" obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI0202 15:15:48.507041 6175 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI0202 15:15:48.507054 6175 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controller-manager-crc in node crc\\\\nI0202 15:15:48.507062 6175 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc after 0 failed attempt(s)\\\\nI0202 15:15:48.507070 6175 default_network_controller.go:776] Recording success event on pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nF0202 15:15:48.507076 6175 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Po\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efaf2216e4c1750139f080a93b4b608737dab2883de5fddfd9a99248d517b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxvkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:49Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.629829 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.629888 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.629905 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.629929 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.629946 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:49Z","lastTransitionTime":"2026-02-02T15:15:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.647561 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8c597" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76cc68c3-9e90-45de-87b2-6ea1b68eadce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ac57831c1c880e64aaa7bedcf999133ee09fc5ecc5124b9d54ecedcc9a13c28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwr7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8c597\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:49Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.669275 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"402a6dc5-6bc7-4d7c-a909-9bb4ca48365d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d5b032497b592363a03f89cb31ba1b551f479004072dddffce540c41190ab81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a3fd02d55a10647428a05a9f53a1429a9827059c523d5deab3c4581c8f1b933\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11600ada13fbc7db1831de385202bc1f041efb656cc58aa49b0f71fe902db9bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fd39f5a05dec7a756dfd519f5df3ef7c77bbb5b8633c5af18a09ff8029cb079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:49Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.690129 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:49Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.711697 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b38b8410bea95465ee58803c4160039bf7c082ca47893680da7469418f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27003b336112760daa2079f4819641850c27fbc8415b63b5094c86195291aca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:49Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.732934 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.732992 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.733011 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.733035 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.733052 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:49Z","lastTransitionTime":"2026-02-02T15:15:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.734927 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jh6vk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76a05d6-d9c7-4183-ad90-b3131f225841\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f2659e2a914d3191daddbfb5ad109e487f03dabb559d324244e127add7d41d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jh6vk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:49Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.753491 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83d0e810-ea5c-47aa-947b-7865b045bf94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6144eb32f88bc106f7724e48133b568a137c4f3fbbfe30c7fcef405d3ab735a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51318ad71b42f4cc8d376aa50e44580e68da89d83d80defb996329a2ee5c2ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zgsv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:49Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.770678 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5xgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56681968-f264-440e-aa12-d4f81879efe9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e250891b1065a38a448aa6abd300b37543e819e87f006b03cc37983cb643982a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdjjl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5xgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:49Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.788624 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7r9fd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4dfd44b-b363-4162-9f43-9a8dc4a3b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7r9fd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:49Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.811364 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7082da2-f2e7-4b73-ba35-504e5cd4c920\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd383d69b9b27d52fa8672ff6e0294a74876d676973bdc0ca95125fef5f2d085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037e379a109ce230098e3ca223fd1dea3149de3ebe601f56e2bb295f3093a3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6fdd19dac506f2d5d7eedc44ec2cda92a9381b73e91a45d949e4fcde4dfc283\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13965f5abcfb0a36c33310518955f192673e630a1aaba94b49400026de5bc2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://992f4198192cbe4b1c89d575aa4a20c29c1115fd5eaa8986c6ed70c1d6b143c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0202 15:15:31.010033 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 15:15:31.011243 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3543528698/tls.crt::/tmp/serving-cert-3543528698/tls.key\\\\\\\"\\\\nI0202 15:15:36.416658 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 15:15:36.428087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 15:15:36.428134 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 15:15:36.428243 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 15:15:36.428280 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 15:15:36.439678 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 15:15:36.439710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439717 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439722 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 15:15:36.439726 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 15:15:36.439729 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 15:15:36.439732 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 15:15:36.439753 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 15:15:36.450716 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://226f5949820d4f498dbf6fa251c6edf1c6356900bd946e6afae2c29042b1c835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:49Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.833662 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da08f9aecc070b31df2ddc1be7c6eb2ad8ef68a931e57ca2e5c5302f4e75751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:49Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.836705 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.836788 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.836815 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.836849 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.836872 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:49Z","lastTransitionTime":"2026-02-02T15:15:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.854787 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:49Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.873216 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca57dabf1d1ee5c6fc29291b1be687ca7e97c72650084b02067057fccac41a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:49Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.893152 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:49Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.940107 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.940203 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.940223 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.940248 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:49 crc kubenswrapper[4733]: I0202 15:15:49.940268 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:49Z","lastTransitionTime":"2026-02-02T15:15:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.042693 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.042765 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.042776 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.042799 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.042814 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:50Z","lastTransitionTime":"2026-02-02T15:15:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.145692 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.145755 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.145772 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.145797 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.145817 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:50Z","lastTransitionTime":"2026-02-02T15:15:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.211713 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 02:41:16.939878758 +0000 UTC Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.248626 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.248675 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.248697 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.248720 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.248736 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:50Z","lastTransitionTime":"2026-02-02T15:15:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.351697 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.351769 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.351786 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.351810 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.351828 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:50Z","lastTransitionTime":"2026-02-02T15:15:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.356826 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hbm26"] Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.365957 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hbm26" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.370673 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.372729 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.387188 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hbm26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45553dc5-54ea-4de9-914d-55b8fee477f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d64j4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d64j4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hbm26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:50Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.419461 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8206eca4-7d8b-4b59-a4c1-24631dd8cd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41d4a7bdc21f00b5039385d7227d8cb0f4ea3051d3b5065c71a49cb0bfad26a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c11b1859bf3bc0ec4d5e6a58d9becc1ac6333a1ed9820d83d634f1e93221c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c08b087d85be9bb83007c7e38a5de9182a5091716fb8a2ca5ded1b8191bebc40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fd4969f0f16c9aa44208d2bf578f77d9d50914a8cb1764507175aa2eb58e733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59175c8e5a7f3536168a69fd944bf3d536bcd7281eee96002343b7836a54401b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:50Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.420260 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/45553dc5-54ea-4de9-914d-55b8fee477f8-env-overrides\") pod \"ovnkube-control-plane-749d76644c-hbm26\" (UID: \"45553dc5-54ea-4de9-914d-55b8fee477f8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hbm26" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.420326 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/45553dc5-54ea-4de9-914d-55b8fee477f8-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-hbm26\" (UID: \"45553dc5-54ea-4de9-914d-55b8fee477f8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hbm26" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.420398 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/45553dc5-54ea-4de9-914d-55b8fee477f8-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-hbm26\" (UID: \"45553dc5-54ea-4de9-914d-55b8fee477f8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hbm26" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.420484 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d64j4\" (UniqueName: \"kubernetes.io/projected/45553dc5-54ea-4de9-914d-55b8fee477f8-kube-api-access-d64j4\") pod \"ovnkube-control-plane-749d76644c-hbm26\" (UID: \"45553dc5-54ea-4de9-914d-55b8fee477f8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hbm26" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.444946 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nr6ps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e43cd902-791a-4042-bc08-cf03e075be32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2a89a5d1fc43696f33d313aee5dca316a249fdef6a23ff655e31f14cc449cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07e56782051c93b2a320aa539a42764b97647237d228bed2b6af48cee987c60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07e56782051c93b2a320aa539a42764b97647237d228bed2b6af48cee987c60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce850ac7bd4c2444b65d87232114d55b78d11c6f973537c615900092a45d55e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce850ac7bd4c2444b65d87232114d55b78d11c6f973537c615900092a45d55e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nr6ps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:50Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.455134 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.455211 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.455231 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.455254 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.455271 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:50Z","lastTransitionTime":"2026-02-02T15:15:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.483985 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5a81ada-e393-499a-aac6-7b40349958bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4acf2398e2e19bd54f482cb2273c54738e61627187860bb0b6daadddc2565b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831d85055203f153381eb51d98af0cd2e4c4a0309f837545ab3d855a680133e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1380122d1694c1286f1102e6396c6504b91f63c6eb157689dfc0ea7188c17f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5835cb53385359e9c58eaa5d4da2569c313e0fc265ad79a6c124e8cecb1a734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd6dc4bf5b9fd31bfbee6d9340fe981c0c5fca03a79cef9316340ff5a8b3d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db1db75e392ebea8b387b85cabc4276d553bf2c06694cf7f294a383909f32826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58d0beeefc5a719f002da8a57bb8f5209f8e9e527159fcfbf77c28cf1a30fdf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd866027caf308b856a71578e47f085bdfb2acef5983d1af659102fb202c6bb0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T15:15:47Z\\\",\\\"message\\\":\\\"ub.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0202 15:15:47.138308 6058 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0202 15:15:47.138443 6058 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0202 15:15:47.138624 6058 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0202 15:15:47.138629 6058 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0202 15:15:47.138693 6058 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0202 15:15:47.139237 6058 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0202 15:15:47.139373 6058 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58d0beeefc5a719f002da8a57bb8f5209f8e9e527159fcfbf77c28cf1a30fdf6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T15:15:48Z\\\",\\\"message\\\":\\\" obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI0202 15:15:48.507041 6175 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI0202 15:15:48.507054 6175 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controller-manager-crc in node crc\\\\nI0202 15:15:48.507062 6175 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc after 0 failed attempt(s)\\\\nI0202 15:15:48.507070 6175 default_network_controller.go:776] Recording success event on pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nF0202 15:15:48.507076 6175 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Po\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efaf2216e4c1750139f080a93b4b608737dab2883de5fddfd9a99248d517b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxvkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:50Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.502151 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8c597" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76cc68c3-9e90-45de-87b2-6ea1b68eadce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ac57831c1c880e64aaa7bedcf999133ee09fc5ecc5124b9d54ecedcc9a13c28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwr7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8c597\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:50Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.516535 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"402a6dc5-6bc7-4d7c-a909-9bb4ca48365d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d5b032497b592363a03f89cb31ba1b551f479004072dddffce540c41190ab81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a3fd02d55a10647428a05a9f53a1429a9827059c523d5deab3c4581c8f1b933\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11600ada13fbc7db1831de385202bc1f041efb656cc58aa49b0f71fe902db9bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fd39f5a05dec7a756dfd519f5df3ef7c77bbb5b8633c5af18a09ff8029cb079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:50Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.521937 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/45553dc5-54ea-4de9-914d-55b8fee477f8-env-overrides\") pod \"ovnkube-control-plane-749d76644c-hbm26\" (UID: \"45553dc5-54ea-4de9-914d-55b8fee477f8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hbm26" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.521996 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/45553dc5-54ea-4de9-914d-55b8fee477f8-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-hbm26\" (UID: \"45553dc5-54ea-4de9-914d-55b8fee477f8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hbm26" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.522604 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/45553dc5-54ea-4de9-914d-55b8fee477f8-env-overrides\") pod \"ovnkube-control-plane-749d76644c-hbm26\" (UID: \"45553dc5-54ea-4de9-914d-55b8fee477f8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hbm26" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.522277 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/45553dc5-54ea-4de9-914d-55b8fee477f8-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-hbm26\" (UID: \"45553dc5-54ea-4de9-914d-55b8fee477f8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hbm26" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.523310 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d64j4\" (UniqueName: \"kubernetes.io/projected/45553dc5-54ea-4de9-914d-55b8fee477f8-kube-api-access-d64j4\") pod \"ovnkube-control-plane-749d76644c-hbm26\" (UID: \"45553dc5-54ea-4de9-914d-55b8fee477f8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hbm26" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.523396 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/45553dc5-54ea-4de9-914d-55b8fee477f8-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-hbm26\" (UID: \"45553dc5-54ea-4de9-914d-55b8fee477f8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hbm26" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.529455 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/45553dc5-54ea-4de9-914d-55b8fee477f8-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-hbm26\" (UID: \"45553dc5-54ea-4de9-914d-55b8fee477f8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hbm26" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.530236 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:50Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.538549 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxvkf_f5a81ada-e393-499a-aac6-7b40349958bb/ovnkube-controller/1.log" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.543759 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d64j4\" (UniqueName: \"kubernetes.io/projected/45553dc5-54ea-4de9-914d-55b8fee477f8-kube-api-access-d64j4\") pod \"ovnkube-control-plane-749d76644c-hbm26\" (UID: \"45553dc5-54ea-4de9-914d-55b8fee477f8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hbm26" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.544197 4733 scope.go:117] "RemoveContainer" containerID="58d0beeefc5a719f002da8a57bb8f5209f8e9e527159fcfbf77c28cf1a30fdf6" Feb 02 15:15:50 crc kubenswrapper[4733]: E0202 15:15:50.544518 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qxvkf_openshift-ovn-kubernetes(f5a81ada-e393-499a-aac6-7b40349958bb)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.554117 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b38b8410bea95465ee58803c4160039bf7c082ca47893680da7469418f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27003b336112760daa2079f4819641850c27fbc8415b63b5094c86195291aca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:50Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.557615 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.557671 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.557691 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.557715 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.557735 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:50Z","lastTransitionTime":"2026-02-02T15:15:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.572273 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jh6vk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76a05d6-d9c7-4183-ad90-b3131f225841\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f2659e2a914d3191daddbfb5ad109e487f03dabb559d324244e127add7d41d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jh6vk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:50Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.587005 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83d0e810-ea5c-47aa-947b-7865b045bf94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6144eb32f88bc106f7724e48133b568a137c4f3fbbfe30c7fcef405d3ab735a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51318ad71b42f4cc8d376aa50e44580e68da89d83d80defb996329a2ee5c2ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zgsv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:50Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.602877 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca57dabf1d1ee5c6fc29291b1be687ca7e97c72650084b02067057fccac41a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:50Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.619669 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:50Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.635756 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5xgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56681968-f264-440e-aa12-d4f81879efe9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e250891b1065a38a448aa6abd300b37543e819e87f006b03cc37983cb643982a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdjjl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5xgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:50Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.653347 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7r9fd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4dfd44b-b363-4162-9f43-9a8dc4a3b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7r9fd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:50Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.659935 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.659986 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.660003 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.660027 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.660044 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:50Z","lastTransitionTime":"2026-02-02T15:15:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.674002 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7082da2-f2e7-4b73-ba35-504e5cd4c920\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd383d69b9b27d52fa8672ff6e0294a74876d676973bdc0ca95125fef5f2d085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037e379a109ce230098e3ca223fd1dea3149de3ebe601f56e2bb295f3093a3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6fdd19dac506f2d5d7eedc44ec2cda92a9381b73e91a45d949e4fcde4dfc283\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13965f5abcfb0a36c33310518955f192673e630a1aaba94b49400026de5bc2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://992f4198192cbe4b1c89d575aa4a20c29c1115fd5eaa8986c6ed70c1d6b143c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0202 15:15:31.010033 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 15:15:31.011243 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3543528698/tls.crt::/tmp/serving-cert-3543528698/tls.key\\\\\\\"\\\\nI0202 15:15:36.416658 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 15:15:36.428087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 15:15:36.428134 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 15:15:36.428243 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 15:15:36.428280 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 15:15:36.439678 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 15:15:36.439710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439717 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439722 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 15:15:36.439726 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 15:15:36.439729 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 15:15:36.439732 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 15:15:36.439753 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 15:15:36.450716 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://226f5949820d4f498dbf6fa251c6edf1c6356900bd946e6afae2c29042b1c835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:50Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.686516 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hbm26" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.691263 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da08f9aecc070b31df2ddc1be7c6eb2ad8ef68a931e57ca2e5c5302f4e75751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:50Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:50 crc kubenswrapper[4733]: W0202 15:15:50.700379 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod45553dc5_54ea_4de9_914d_55b8fee477f8.slice/crio-4a7bc65997f17b98a1189ea8ad0d67abb9e4a06a5aa761130a65de644e70f6fb WatchSource:0}: Error finding container 4a7bc65997f17b98a1189ea8ad0d67abb9e4a06a5aa761130a65de644e70f6fb: Status 404 returned error can't find the container with id 4a7bc65997f17b98a1189ea8ad0d67abb9e4a06a5aa761130a65de644e70f6fb Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.713894 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:50Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.735044 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"402a6dc5-6bc7-4d7c-a909-9bb4ca48365d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d5b032497b592363a03f89cb31ba1b551f479004072dddffce540c41190ab81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a3fd02d55a10647428a05a9f53a1429a9827059c523d5deab3c4581c8f1b933\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11600ada13fbc7db1831de385202bc1f041efb656cc58aa49b0f71fe902db9bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fd39f5a05dec7a756dfd519f5df3ef7c77bbb5b8633c5af18a09ff8029cb079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:50Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.756438 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:50Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.762620 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.762713 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.762731 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.762755 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.762771 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:50Z","lastTransitionTime":"2026-02-02T15:15:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.777029 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b38b8410bea95465ee58803c4160039bf7c082ca47893680da7469418f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27003b336112760daa2079f4819641850c27fbc8415b63b5094c86195291aca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:50Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.796422 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jh6vk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76a05d6-d9c7-4183-ad90-b3131f225841\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f2659e2a914d3191daddbfb5ad109e487f03dabb559d324244e127add7d41d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jh6vk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:50Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.808849 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83d0e810-ea5c-47aa-947b-7865b045bf94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6144eb32f88bc106f7724e48133b568a137c4f3fbbfe30c7fcef405d3ab735a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51318ad71b42f4cc8d376aa50e44580e68da89d83d80defb996329a2ee5c2ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zgsv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:50Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.827953 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7082da2-f2e7-4b73-ba35-504e5cd4c920\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd383d69b9b27d52fa8672ff6e0294a74876d676973bdc0ca95125fef5f2d085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037e379a109ce230098e3ca223fd1dea3149de3ebe601f56e2bb295f3093a3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6fdd19dac506f2d5d7eedc44ec2cda92a9381b73e91a45d949e4fcde4dfc283\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13965f5abcfb0a36c33310518955f192673e630a1aaba94b49400026de5bc2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://992f4198192cbe4b1c89d575aa4a20c29c1115fd5eaa8986c6ed70c1d6b143c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0202 15:15:31.010033 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 15:15:31.011243 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3543528698/tls.crt::/tmp/serving-cert-3543528698/tls.key\\\\\\\"\\\\nI0202 15:15:36.416658 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 15:15:36.428087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 15:15:36.428134 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 15:15:36.428243 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 15:15:36.428280 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 15:15:36.439678 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 15:15:36.439710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439717 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439722 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 15:15:36.439726 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 15:15:36.439729 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 15:15:36.439732 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 15:15:36.439753 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 15:15:36.450716 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://226f5949820d4f498dbf6fa251c6edf1c6356900bd946e6afae2c29042b1c835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:50Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.849398 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da08f9aecc070b31df2ddc1be7c6eb2ad8ef68a931e57ca2e5c5302f4e75751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:50Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.866778 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:50Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.867620 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.867661 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.867674 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.867694 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.867706 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:50Z","lastTransitionTime":"2026-02-02T15:15:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.883324 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca57dabf1d1ee5c6fc29291b1be687ca7e97c72650084b02067057fccac41a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:50Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.903784 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:50Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.919217 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5xgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56681968-f264-440e-aa12-d4f81879efe9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e250891b1065a38a448aa6abd300b37543e819e87f006b03cc37983cb643982a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdjjl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5xgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:50Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.933519 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7r9fd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4dfd44b-b363-4162-9f43-9a8dc4a3b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7r9fd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:50Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.959534 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8206eca4-7d8b-4b59-a4c1-24631dd8cd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41d4a7bdc21f00b5039385d7227d8cb0f4ea3051d3b5065c71a49cb0bfad26a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c11b1859bf3bc0ec4d5e6a58d9becc1ac6333a1ed9820d83d634f1e93221c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c08b087d85be9bb83007c7e38a5de9182a5091716fb8a2ca5ded1b8191bebc40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fd4969f0f16c9aa44208d2bf578f77d9d50914a8cb1764507175aa2eb58e733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59175c8e5a7f3536168a69fd944bf3d536bcd7281eee96002343b7836a54401b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:50Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.971644 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.971687 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.971704 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.971723 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.971738 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:50Z","lastTransitionTime":"2026-02-02T15:15:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:50 crc kubenswrapper[4733]: I0202 15:15:50.983316 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nr6ps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e43cd902-791a-4042-bc08-cf03e075be32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2a89a5d1fc43696f33d313aee5dca316a249fdef6a23ff655e31f14cc449cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07e56782051c93b2a320aa539a42764b97647237d228bed2b6af48cee987c60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07e56782051c93b2a320aa539a42764b97647237d228bed2b6af48cee987c60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce850ac7bd4c2444b65d87232114d55b78d11c6f973537c615900092a45d55e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce850ac7bd4c2444b65d87232114d55b78d11c6f973537c615900092a45d55e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nr6ps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:50Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.023795 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5a81ada-e393-499a-aac6-7b40349958bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4acf2398e2e19bd54f482cb2273c54738e61627187860bb0b6daadddc2565b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831d85055203f153381eb51d98af0cd2e4c4a0309f837545ab3d855a680133e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1380122d1694c1286f1102e6396c6504b91f63c6eb157689dfc0ea7188c17f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5835cb53385359e9c58eaa5d4da2569c313e0fc265ad79a6c124e8cecb1a734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd6dc4bf5b9fd31bfbee6d9340fe981c0c5fca03a79cef9316340ff5a8b3d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db1db75e392ebea8b387b85cabc4276d553bf2c06694cf7f294a383909f32826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58d0beeefc5a719f002da8a57bb8f5209f8e9e527159fcfbf77c28cf1a30fdf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58d0beeefc5a719f002da8a57bb8f5209f8e9e527159fcfbf77c28cf1a30fdf6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T15:15:48Z\\\",\\\"message\\\":\\\" obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI0202 15:15:48.507041 6175 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI0202 15:15:48.507054 6175 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controller-manager-crc in node crc\\\\nI0202 15:15:48.507062 6175 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc after 0 failed attempt(s)\\\\nI0202 15:15:48.507070 6175 default_network_controller.go:776] Recording success event on pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nF0202 15:15:48.507076 6175 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Po\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qxvkf_openshift-ovn-kubernetes(f5a81ada-e393-499a-aac6-7b40349958bb)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efaf2216e4c1750139f080a93b4b608737dab2883de5fddfd9a99248d517b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxvkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:51Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.039122 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8c597" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76cc68c3-9e90-45de-87b2-6ea1b68eadce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ac57831c1c880e64aaa7bedcf999133ee09fc5ecc5124b9d54ecedcc9a13c28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwr7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8c597\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:51Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.053976 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hbm26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45553dc5-54ea-4de9-914d-55b8fee477f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d64j4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d64j4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hbm26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:51Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.074844 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.074891 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.074903 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.074923 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.074935 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:51Z","lastTransitionTime":"2026-02-02T15:15:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.178208 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.178559 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.178726 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.178863 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.179000 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:51Z","lastTransitionTime":"2026-02-02T15:15:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.212582 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-11 09:58:39.307352861 +0000 UTC Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.254190 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:15:51 crc kubenswrapper[4733]: E0202 15:15:51.254336 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.254380 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.254447 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:15:51 crc kubenswrapper[4733]: E0202 15:15:51.254485 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.254494 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:15:51 crc kubenswrapper[4733]: E0202 15:15:51.254646 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:15:51 crc kubenswrapper[4733]: E0202 15:15:51.254737 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.280963 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.281010 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.281025 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.281045 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.281061 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:51Z","lastTransitionTime":"2026-02-02T15:15:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.383445 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.383497 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.383509 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.383529 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.383542 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:51Z","lastTransitionTime":"2026-02-02T15:15:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.486958 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.487018 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.487041 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.487071 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.487093 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:51Z","lastTransitionTime":"2026-02-02T15:15:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.549923 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hbm26" event={"ID":"45553dc5-54ea-4de9-914d-55b8fee477f8","Type":"ContainerStarted","Data":"f76a1f655f871deb7e1364fcc5dbf6b564c4c386300bb49cb75f83518b64286e"} Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.549980 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hbm26" event={"ID":"45553dc5-54ea-4de9-914d-55b8fee477f8","Type":"ContainerStarted","Data":"5d112b1ea41c3d30a7cfa46d17db6aed0999eb0baaaf8a2f21c741d805a21451"} Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.550001 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hbm26" event={"ID":"45553dc5-54ea-4de9-914d-55b8fee477f8","Type":"ContainerStarted","Data":"4a7bc65997f17b98a1189ea8ad0d67abb9e4a06a5aa761130a65de644e70f6fb"} Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.571300 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:51Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.586904 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5xgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56681968-f264-440e-aa12-d4f81879efe9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e250891b1065a38a448aa6abd300b37543e819e87f006b03cc37983cb643982a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdjjl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5xgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:51Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.590044 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.590125 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.590150 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.590235 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.590263 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:51Z","lastTransitionTime":"2026-02-02T15:15:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.603883 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7r9fd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4dfd44b-b363-4162-9f43-9a8dc4a3b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7r9fd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:51Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.626468 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7082da2-f2e7-4b73-ba35-504e5cd4c920\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd383d69b9b27d52fa8672ff6e0294a74876d676973bdc0ca95125fef5f2d085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037e379a109ce230098e3ca223fd1dea3149de3ebe601f56e2bb295f3093a3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6fdd19dac506f2d5d7eedc44ec2cda92a9381b73e91a45d949e4fcde4dfc283\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13965f5abcfb0a36c33310518955f192673e630a1aaba94b49400026de5bc2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://992f4198192cbe4b1c89d575aa4a20c29c1115fd5eaa8986c6ed70c1d6b143c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0202 15:15:31.010033 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 15:15:31.011243 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3543528698/tls.crt::/tmp/serving-cert-3543528698/tls.key\\\\\\\"\\\\nI0202 15:15:36.416658 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 15:15:36.428087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 15:15:36.428134 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 15:15:36.428243 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 15:15:36.428280 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 15:15:36.439678 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 15:15:36.439710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439717 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439722 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 15:15:36.439726 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 15:15:36.439729 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 15:15:36.439732 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 15:15:36.439753 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 15:15:36.450716 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://226f5949820d4f498dbf6fa251c6edf1c6356900bd946e6afae2c29042b1c835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:51Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.648983 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da08f9aecc070b31df2ddc1be7c6eb2ad8ef68a931e57ca2e5c5302f4e75751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:51Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.668013 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:51Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.686923 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca57dabf1d1ee5c6fc29291b1be687ca7e97c72650084b02067057fccac41a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:51Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.693044 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.693107 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.693125 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.693150 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.693195 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:51Z","lastTransitionTime":"2026-02-02T15:15:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.721501 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8206eca4-7d8b-4b59-a4c1-24631dd8cd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41d4a7bdc21f00b5039385d7227d8cb0f4ea3051d3b5065c71a49cb0bfad26a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c11b1859bf3bc0ec4d5e6a58d9becc1ac6333a1ed9820d83d634f1e93221c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c08b087d85be9bb83007c7e38a5de9182a5091716fb8a2ca5ded1b8191bebc40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fd4969f0f16c9aa44208d2bf578f77d9d50914a8cb1764507175aa2eb58e733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59175c8e5a7f3536168a69fd944bf3d536bcd7281eee96002343b7836a54401b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:51Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.746055 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nr6ps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e43cd902-791a-4042-bc08-cf03e075be32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2a89a5d1fc43696f33d313aee5dca316a249fdef6a23ff655e31f14cc449cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07e56782051c93b2a320aa539a42764b97647237d228bed2b6af48cee987c60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07e56782051c93b2a320aa539a42764b97647237d228bed2b6af48cee987c60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce850ac7bd4c2444b65d87232114d55b78d11c6f973537c615900092a45d55e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce850ac7bd4c2444b65d87232114d55b78d11c6f973537c615900092a45d55e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nr6ps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:51Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.779589 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5a81ada-e393-499a-aac6-7b40349958bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4acf2398e2e19bd54f482cb2273c54738e61627187860bb0b6daadddc2565b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831d85055203f153381eb51d98af0cd2e4c4a0309f837545ab3d855a680133e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1380122d1694c1286f1102e6396c6504b91f63c6eb157689dfc0ea7188c17f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5835cb53385359e9c58eaa5d4da2569c313e0fc265ad79a6c124e8cecb1a734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd6dc4bf5b9fd31bfbee6d9340fe981c0c5fca03a79cef9316340ff5a8b3d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db1db75e392ebea8b387b85cabc4276d553bf2c06694cf7f294a383909f32826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58d0beeefc5a719f002da8a57bb8f5209f8e9e527159fcfbf77c28cf1a30fdf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58d0beeefc5a719f002da8a57bb8f5209f8e9e527159fcfbf77c28cf1a30fdf6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T15:15:48Z\\\",\\\"message\\\":\\\" obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI0202 15:15:48.507041 6175 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI0202 15:15:48.507054 6175 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controller-manager-crc in node crc\\\\nI0202 15:15:48.507062 6175 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc after 0 failed attempt(s)\\\\nI0202 15:15:48.507070 6175 default_network_controller.go:776] Recording success event on pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nF0202 15:15:48.507076 6175 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Po\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qxvkf_openshift-ovn-kubernetes(f5a81ada-e393-499a-aac6-7b40349958bb)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efaf2216e4c1750139f080a93b4b608737dab2883de5fddfd9a99248d517b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxvkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:51Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.796323 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.796382 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.796403 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.796429 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.796450 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:51Z","lastTransitionTime":"2026-02-02T15:15:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.796564 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8c597" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76cc68c3-9e90-45de-87b2-6ea1b68eadce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ac57831c1c880e64aaa7bedcf999133ee09fc5ecc5124b9d54ecedcc9a13c28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwr7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8c597\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:51Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.814472 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hbm26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45553dc5-54ea-4de9-914d-55b8fee477f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d112b1ea41c3d30a7cfa46d17db6aed0999eb0baaaf8a2f21c741d805a21451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d64j4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f76a1f655f871deb7e1364fcc5dbf6b564c4c386300bb49cb75f83518b64286e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d64j4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hbm26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:51Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.833126 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"402a6dc5-6bc7-4d7c-a909-9bb4ca48365d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d5b032497b592363a03f89cb31ba1b551f479004072dddffce540c41190ab81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a3fd02d55a10647428a05a9f53a1429a9827059c523d5deab3c4581c8f1b933\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11600ada13fbc7db1831de385202bc1f041efb656cc58aa49b0f71fe902db9bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fd39f5a05dec7a756dfd519f5df3ef7c77bbb5b8633c5af18a09ff8029cb079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:51Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.853390 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:51Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.872829 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b38b8410bea95465ee58803c4160039bf7c082ca47893680da7469418f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27003b336112760daa2079f4819641850c27fbc8415b63b5094c86195291aca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:51Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.895029 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jh6vk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76a05d6-d9c7-4183-ad90-b3131f225841\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f2659e2a914d3191daddbfb5ad109e487f03dabb559d324244e127add7d41d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jh6vk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:51Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.899265 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.899323 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.899340 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.899364 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.899382 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:51Z","lastTransitionTime":"2026-02-02T15:15:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:51 crc kubenswrapper[4733]: I0202 15:15:51.912037 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83d0e810-ea5c-47aa-947b-7865b045bf94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6144eb32f88bc106f7724e48133b568a137c4f3fbbfe30c7fcef405d3ab735a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51318ad71b42f4cc8d376aa50e44580e68da89d83d80defb996329a2ee5c2ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zgsv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:51Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.002136 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.002240 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.002259 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.002307 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.002325 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:52Z","lastTransitionTime":"2026-02-02T15:15:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.105088 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.105137 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.105148 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.105183 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.105195 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:52Z","lastTransitionTime":"2026-02-02T15:15:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.208718 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.208771 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.208787 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.208809 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.208827 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:52Z","lastTransitionTime":"2026-02-02T15:15:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.212971 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-09 14:34:09.158610953 +0000 UTC Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.311379 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.311438 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.311455 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.311480 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.311496 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:52Z","lastTransitionTime":"2026-02-02T15:15:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.413728 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.413762 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.413774 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.413793 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.413806 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:52Z","lastTransitionTime":"2026-02-02T15:15:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.517043 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.517113 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.517134 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.517203 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.517226 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:52Z","lastTransitionTime":"2026-02-02T15:15:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.620817 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.621933 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.621996 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.622034 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.622058 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:52Z","lastTransitionTime":"2026-02-02T15:15:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.725409 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.725481 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.725493 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.725540 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.725555 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:52Z","lastTransitionTime":"2026-02-02T15:15:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.828093 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.828271 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.828292 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.828318 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.828336 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:52Z","lastTransitionTime":"2026-02-02T15:15:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.848945 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 15:15:52 crc kubenswrapper[4733]: E0202 15:15:52.849127 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 15:16:08.849091667 +0000 UTC m=+52.300553065 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.849441 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:15:52 crc kubenswrapper[4733]: E0202 15:15:52.849674 4733 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 02 15:15:52 crc kubenswrapper[4733]: E0202 15:15:52.849760 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-02 15:16:08.849744014 +0000 UTC m=+52.301205412 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.930781 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.930838 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.930854 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.930877 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.930894 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:52Z","lastTransitionTime":"2026-02-02T15:15:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.951591 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.951716 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d4dfd44b-b363-4162-9f43-9a8dc4a3b110-metrics-certs\") pod \"network-metrics-daemon-7r9fd\" (UID: \"d4dfd44b-b363-4162-9f43-9a8dc4a3b110\") " pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:15:52 crc kubenswrapper[4733]: E0202 15:15:52.951738 4733 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.951781 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:15:52 crc kubenswrapper[4733]: E0202 15:15:52.951862 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-02 15:16:08.951827202 +0000 UTC m=+52.403288600 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 02 15:15:52 crc kubenswrapper[4733]: E0202 15:15:52.951872 4733 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 02 15:15:52 crc kubenswrapper[4733]: I0202 15:15:52.951920 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:15:52 crc kubenswrapper[4733]: E0202 15:15:52.951948 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d4dfd44b-b363-4162-9f43-9a8dc4a3b110-metrics-certs podName:d4dfd44b-b363-4162-9f43-9a8dc4a3b110 nodeName:}" failed. No retries permitted until 2026-02-02 15:16:08.951926365 +0000 UTC m=+52.403387813 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d4dfd44b-b363-4162-9f43-9a8dc4a3b110-metrics-certs") pod "network-metrics-daemon-7r9fd" (UID: "d4dfd44b-b363-4162-9f43-9a8dc4a3b110") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 02 15:15:52 crc kubenswrapper[4733]: E0202 15:15:52.951975 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 02 15:15:52 crc kubenswrapper[4733]: E0202 15:15:52.952006 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 02 15:15:52 crc kubenswrapper[4733]: E0202 15:15:52.952030 4733 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 15:15:52 crc kubenswrapper[4733]: E0202 15:15:52.952106 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 02 15:15:52 crc kubenswrapper[4733]: E0202 15:15:52.952142 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-02 15:16:08.95211048 +0000 UTC m=+52.403571888 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 15:15:52 crc kubenswrapper[4733]: E0202 15:15:52.952143 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 02 15:15:52 crc kubenswrapper[4733]: E0202 15:15:52.952229 4733 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 15:15:52 crc kubenswrapper[4733]: E0202 15:15:52.952295 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-02 15:16:08.952274914 +0000 UTC m=+52.403736362 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.033939 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.034002 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.034025 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.034054 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.034076 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:53Z","lastTransitionTime":"2026-02-02T15:15:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.137088 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.137436 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.137599 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.137745 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.137881 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:53Z","lastTransitionTime":"2026-02-02T15:15:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.213301 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-06 22:20:24.755988006 +0000 UTC Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.240603 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.240863 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.241016 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.241157 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.241356 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:53Z","lastTransitionTime":"2026-02-02T15:15:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.254937 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.255032 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.255098 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:15:53 crc kubenswrapper[4733]: E0202 15:15:53.255228 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.255277 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:15:53 crc kubenswrapper[4733]: E0202 15:15:53.255486 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:15:53 crc kubenswrapper[4733]: E0202 15:15:53.255637 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:15:53 crc kubenswrapper[4733]: E0202 15:15:53.255777 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.344335 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.344399 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.344416 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.344442 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.344459 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:53Z","lastTransitionTime":"2026-02-02T15:15:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.447258 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.447336 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.447354 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.447381 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.447399 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:53Z","lastTransitionTime":"2026-02-02T15:15:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.549974 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.550028 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.550044 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.550069 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.550086 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:53Z","lastTransitionTime":"2026-02-02T15:15:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.652848 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.652887 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.652899 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.652916 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.652929 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:53Z","lastTransitionTime":"2026-02-02T15:15:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.756075 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.756248 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.756363 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.756392 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.756409 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:53Z","lastTransitionTime":"2026-02-02T15:15:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.859346 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.859490 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.859513 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.859538 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.859560 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:53Z","lastTransitionTime":"2026-02-02T15:15:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.962760 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.963098 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.963274 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.963411 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:53 crc kubenswrapper[4733]: I0202 15:15:53.963596 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:53Z","lastTransitionTime":"2026-02-02T15:15:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:54 crc kubenswrapper[4733]: I0202 15:15:54.066956 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:54 crc kubenswrapper[4733]: I0202 15:15:54.067364 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:54 crc kubenswrapper[4733]: I0202 15:15:54.067489 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:54 crc kubenswrapper[4733]: I0202 15:15:54.067602 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:54 crc kubenswrapper[4733]: I0202 15:15:54.067730 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:54Z","lastTransitionTime":"2026-02-02T15:15:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:54 crc kubenswrapper[4733]: I0202 15:15:54.171268 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:54 crc kubenswrapper[4733]: I0202 15:15:54.171335 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:54 crc kubenswrapper[4733]: I0202 15:15:54.171357 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:54 crc kubenswrapper[4733]: I0202 15:15:54.171382 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:54 crc kubenswrapper[4733]: I0202 15:15:54.171399 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:54Z","lastTransitionTime":"2026-02-02T15:15:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:54 crc kubenswrapper[4733]: I0202 15:15:54.213497 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 23:04:39.200148312 +0000 UTC Feb 02 15:15:54 crc kubenswrapper[4733]: I0202 15:15:54.274392 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:54 crc kubenswrapper[4733]: I0202 15:15:54.274428 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:54 crc kubenswrapper[4733]: I0202 15:15:54.274437 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:54 crc kubenswrapper[4733]: I0202 15:15:54.274451 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:54 crc kubenswrapper[4733]: I0202 15:15:54.274460 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:54Z","lastTransitionTime":"2026-02-02T15:15:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:54 crc kubenswrapper[4733]: I0202 15:15:54.377700 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:54 crc kubenswrapper[4733]: I0202 15:15:54.377780 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:54 crc kubenswrapper[4733]: I0202 15:15:54.377805 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:54 crc kubenswrapper[4733]: I0202 15:15:54.377836 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:54 crc kubenswrapper[4733]: I0202 15:15:54.377861 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:54Z","lastTransitionTime":"2026-02-02T15:15:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:54 crc kubenswrapper[4733]: I0202 15:15:54.480616 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:54 crc kubenswrapper[4733]: I0202 15:15:54.480682 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:54 crc kubenswrapper[4733]: I0202 15:15:54.480700 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:54 crc kubenswrapper[4733]: I0202 15:15:54.480725 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:54 crc kubenswrapper[4733]: I0202 15:15:54.480746 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:54Z","lastTransitionTime":"2026-02-02T15:15:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:54 crc kubenswrapper[4733]: I0202 15:15:54.583519 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:54 crc kubenswrapper[4733]: I0202 15:15:54.583555 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:54 crc kubenswrapper[4733]: I0202 15:15:54.583564 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:54 crc kubenswrapper[4733]: I0202 15:15:54.583580 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:54 crc kubenswrapper[4733]: I0202 15:15:54.583590 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:54Z","lastTransitionTime":"2026-02-02T15:15:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:54 crc kubenswrapper[4733]: I0202 15:15:54.687131 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:54 crc kubenswrapper[4733]: I0202 15:15:54.687281 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:54 crc kubenswrapper[4733]: I0202 15:15:54.687321 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:54 crc kubenswrapper[4733]: I0202 15:15:54.687350 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:54 crc kubenswrapper[4733]: I0202 15:15:54.687369 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:54Z","lastTransitionTime":"2026-02-02T15:15:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:54 crc kubenswrapper[4733]: I0202 15:15:54.790707 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:54 crc kubenswrapper[4733]: I0202 15:15:54.790753 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:54 crc kubenswrapper[4733]: I0202 15:15:54.790762 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:54 crc kubenswrapper[4733]: I0202 15:15:54.790780 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:54 crc kubenswrapper[4733]: I0202 15:15:54.790790 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:54Z","lastTransitionTime":"2026-02-02T15:15:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:54 crc kubenswrapper[4733]: I0202 15:15:54.894206 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:54 crc kubenswrapper[4733]: I0202 15:15:54.894242 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:54 crc kubenswrapper[4733]: I0202 15:15:54.894253 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:54 crc kubenswrapper[4733]: I0202 15:15:54.894267 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:54 crc kubenswrapper[4733]: I0202 15:15:54.894277 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:54Z","lastTransitionTime":"2026-02-02T15:15:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:54 crc kubenswrapper[4733]: I0202 15:15:54.997132 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:54 crc kubenswrapper[4733]: I0202 15:15:54.997205 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:54 crc kubenswrapper[4733]: I0202 15:15:54.997219 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:54 crc kubenswrapper[4733]: I0202 15:15:54.997236 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:54 crc kubenswrapper[4733]: I0202 15:15:54.997248 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:54Z","lastTransitionTime":"2026-02-02T15:15:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:55 crc kubenswrapper[4733]: I0202 15:15:55.099887 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:55 crc kubenswrapper[4733]: I0202 15:15:55.099944 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:55 crc kubenswrapper[4733]: I0202 15:15:55.099962 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:55 crc kubenswrapper[4733]: I0202 15:15:55.099984 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:55 crc kubenswrapper[4733]: I0202 15:15:55.100002 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:55Z","lastTransitionTime":"2026-02-02T15:15:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:55 crc kubenswrapper[4733]: I0202 15:15:55.202649 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:55 crc kubenswrapper[4733]: I0202 15:15:55.202725 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:55 crc kubenswrapper[4733]: I0202 15:15:55.202750 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:55 crc kubenswrapper[4733]: I0202 15:15:55.202784 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:55 crc kubenswrapper[4733]: I0202 15:15:55.202807 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:55Z","lastTransitionTime":"2026-02-02T15:15:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:55 crc kubenswrapper[4733]: I0202 15:15:55.214284 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-06 02:54:06.352488882 +0000 UTC Feb 02 15:15:55 crc kubenswrapper[4733]: I0202 15:15:55.254221 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:15:55 crc kubenswrapper[4733]: I0202 15:15:55.254262 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:15:55 crc kubenswrapper[4733]: I0202 15:15:55.254299 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:15:55 crc kubenswrapper[4733]: E0202 15:15:55.254382 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:15:55 crc kubenswrapper[4733]: I0202 15:15:55.254458 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:15:55 crc kubenswrapper[4733]: E0202 15:15:55.254719 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:15:55 crc kubenswrapper[4733]: E0202 15:15:55.254788 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:15:55 crc kubenswrapper[4733]: E0202 15:15:55.254874 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:15:55 crc kubenswrapper[4733]: I0202 15:15:55.305538 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:55 crc kubenswrapper[4733]: I0202 15:15:55.305611 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:55 crc kubenswrapper[4733]: I0202 15:15:55.305635 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:55 crc kubenswrapper[4733]: I0202 15:15:55.305665 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:55 crc kubenswrapper[4733]: I0202 15:15:55.305691 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:55Z","lastTransitionTime":"2026-02-02T15:15:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:55 crc kubenswrapper[4733]: I0202 15:15:55.409349 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:55 crc kubenswrapper[4733]: I0202 15:15:55.409413 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:55 crc kubenswrapper[4733]: I0202 15:15:55.409435 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:55 crc kubenswrapper[4733]: I0202 15:15:55.409463 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:55 crc kubenswrapper[4733]: I0202 15:15:55.409487 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:55Z","lastTransitionTime":"2026-02-02T15:15:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:55 crc kubenswrapper[4733]: I0202 15:15:55.512535 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:55 crc kubenswrapper[4733]: I0202 15:15:55.512609 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:55 crc kubenswrapper[4733]: I0202 15:15:55.512626 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:55 crc kubenswrapper[4733]: I0202 15:15:55.512649 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:55 crc kubenswrapper[4733]: I0202 15:15:55.512668 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:55Z","lastTransitionTime":"2026-02-02T15:15:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:55 crc kubenswrapper[4733]: I0202 15:15:55.615438 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:55 crc kubenswrapper[4733]: I0202 15:15:55.615483 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:55 crc kubenswrapper[4733]: I0202 15:15:55.615500 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:55 crc kubenswrapper[4733]: I0202 15:15:55.615524 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:55 crc kubenswrapper[4733]: I0202 15:15:55.615540 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:55Z","lastTransitionTime":"2026-02-02T15:15:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:55 crc kubenswrapper[4733]: I0202 15:15:55.718245 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:55 crc kubenswrapper[4733]: I0202 15:15:55.718307 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:55 crc kubenswrapper[4733]: I0202 15:15:55.718316 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:55 crc kubenswrapper[4733]: I0202 15:15:55.718333 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:55 crc kubenswrapper[4733]: I0202 15:15:55.718343 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:55Z","lastTransitionTime":"2026-02-02T15:15:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:55 crc kubenswrapper[4733]: I0202 15:15:55.821442 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:55 crc kubenswrapper[4733]: I0202 15:15:55.821492 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:55 crc kubenswrapper[4733]: I0202 15:15:55.821504 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:55 crc kubenswrapper[4733]: I0202 15:15:55.821520 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:55 crc kubenswrapper[4733]: I0202 15:15:55.821532 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:55Z","lastTransitionTime":"2026-02-02T15:15:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:55 crc kubenswrapper[4733]: I0202 15:15:55.924428 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:55 crc kubenswrapper[4733]: I0202 15:15:55.924489 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:55 crc kubenswrapper[4733]: I0202 15:15:55.924514 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:55 crc kubenswrapper[4733]: I0202 15:15:55.924545 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:55 crc kubenswrapper[4733]: I0202 15:15:55.924570 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:55Z","lastTransitionTime":"2026-02-02T15:15:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.027453 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.027486 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.027494 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.027507 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.027516 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:56Z","lastTransitionTime":"2026-02-02T15:15:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.130012 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.130087 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.130112 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.130142 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.130204 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:56Z","lastTransitionTime":"2026-02-02T15:15:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.214536 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 02:34:34.077977141 +0000 UTC Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.232429 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.232472 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.232493 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.232518 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.232535 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:56Z","lastTransitionTime":"2026-02-02T15:15:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.335570 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.335623 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.335643 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.335664 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.335681 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:56Z","lastTransitionTime":"2026-02-02T15:15:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.438784 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.438842 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.438859 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.438890 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.438908 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:56Z","lastTransitionTime":"2026-02-02T15:15:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.541791 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.541847 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.541864 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.541887 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.541903 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:56Z","lastTransitionTime":"2026-02-02T15:15:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.644519 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.644576 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.644594 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.644616 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.644633 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:56Z","lastTransitionTime":"2026-02-02T15:15:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.747538 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.747594 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.747611 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.747641 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.747659 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:56Z","lastTransitionTime":"2026-02-02T15:15:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.850147 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.850249 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.850316 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.850359 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.850380 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:56Z","lastTransitionTime":"2026-02-02T15:15:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.952801 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.952875 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.952931 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.952955 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.952972 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:56Z","lastTransitionTime":"2026-02-02T15:15:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.973646 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.973700 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.973720 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.973746 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.973770 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:56Z","lastTransitionTime":"2026-02-02T15:15:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:56 crc kubenswrapper[4733]: E0202 15:15:56.990436 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd742269-802e-45ec-a88e-17aee17c574e\\\",\\\"systemUUID\\\":\\\"91b64b75-3cd9-4fb9-93f0-dbd3bcdcaec4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:56Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.999053 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.999121 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.999144 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.999203 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:56 crc kubenswrapper[4733]: I0202 15:15:56.999257 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:56Z","lastTransitionTime":"2026-02-02T15:15:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:57 crc kubenswrapper[4733]: E0202 15:15:57.018679 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd742269-802e-45ec-a88e-17aee17c574e\\\",\\\"systemUUID\\\":\\\"91b64b75-3cd9-4fb9-93f0-dbd3bcdcaec4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:57Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.022468 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.022504 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.022515 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.022530 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.022541 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:57Z","lastTransitionTime":"2026-02-02T15:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:57 crc kubenswrapper[4733]: E0202 15:15:57.040684 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd742269-802e-45ec-a88e-17aee17c574e\\\",\\\"systemUUID\\\":\\\"91b64b75-3cd9-4fb9-93f0-dbd3bcdcaec4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:57Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.043655 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.043684 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.043692 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.043706 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.043717 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:57Z","lastTransitionTime":"2026-02-02T15:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:57 crc kubenswrapper[4733]: E0202 15:15:57.056558 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd742269-802e-45ec-a88e-17aee17c574e\\\",\\\"systemUUID\\\":\\\"91b64b75-3cd9-4fb9-93f0-dbd3bcdcaec4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:57Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.060906 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.060994 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.061007 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.061020 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.061051 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:57Z","lastTransitionTime":"2026-02-02T15:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:57 crc kubenswrapper[4733]: E0202 15:15:57.078218 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:15:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd742269-802e-45ec-a88e-17aee17c574e\\\",\\\"systemUUID\\\":\\\"91b64b75-3cd9-4fb9-93f0-dbd3bcdcaec4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:57Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:57 crc kubenswrapper[4733]: E0202 15:15:57.078340 4733 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.079666 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.079730 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.079754 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.079781 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.079801 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:57Z","lastTransitionTime":"2026-02-02T15:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.183395 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.183465 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.183488 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.183516 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.183537 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:57Z","lastTransitionTime":"2026-02-02T15:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.215445 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 03:22:41.728881715 +0000 UTC Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.254907 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.255018 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:15:57 crc kubenswrapper[4733]: E0202 15:15:57.255123 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.254939 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:15:57 crc kubenswrapper[4733]: E0202 15:15:57.255353 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.255393 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:15:57 crc kubenswrapper[4733]: E0202 15:15:57.255467 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:15:57 crc kubenswrapper[4733]: E0202 15:15:57.255843 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.274255 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:57Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.286978 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.287022 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.287030 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.287046 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.287056 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:57Z","lastTransitionTime":"2026-02-02T15:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.289770 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b38b8410bea95465ee58803c4160039bf7c082ca47893680da7469418f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27003b336112760daa2079f4819641850c27fbc8415b63b5094c86195291aca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:57Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.310009 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jh6vk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76a05d6-d9c7-4183-ad90-b3131f225841\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f2659e2a914d3191daddbfb5ad109e487f03dabb559d324244e127add7d41d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jh6vk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:57Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.331339 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83d0e810-ea5c-47aa-947b-7865b045bf94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6144eb32f88bc106f7724e48133b568a137c4f3fbbfe30c7fcef405d3ab735a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51318ad71b42f4cc8d376aa50e44580e68da89d83d80defb996329a2ee5c2ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zgsv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:57Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.348922 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5xgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56681968-f264-440e-aa12-d4f81879efe9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e250891b1065a38a448aa6abd300b37543e819e87f006b03cc37983cb643982a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdjjl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5xgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:57Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.364324 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7r9fd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4dfd44b-b363-4162-9f43-9a8dc4a3b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7r9fd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:57Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.386051 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7082da2-f2e7-4b73-ba35-504e5cd4c920\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd383d69b9b27d52fa8672ff6e0294a74876d676973bdc0ca95125fef5f2d085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037e379a109ce230098e3ca223fd1dea3149de3ebe601f56e2bb295f3093a3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6fdd19dac506f2d5d7eedc44ec2cda92a9381b73e91a45d949e4fcde4dfc283\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13965f5abcfb0a36c33310518955f192673e630a1aaba94b49400026de5bc2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://992f4198192cbe4b1c89d575aa4a20c29c1115fd5eaa8986c6ed70c1d6b143c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0202 15:15:31.010033 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 15:15:31.011243 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3543528698/tls.crt::/tmp/serving-cert-3543528698/tls.key\\\\\\\"\\\\nI0202 15:15:36.416658 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 15:15:36.428087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 15:15:36.428134 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 15:15:36.428243 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 15:15:36.428280 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 15:15:36.439678 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 15:15:36.439710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439717 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439722 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 15:15:36.439726 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 15:15:36.439729 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 15:15:36.439732 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 15:15:36.439753 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 15:15:36.450716 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://226f5949820d4f498dbf6fa251c6edf1c6356900bd946e6afae2c29042b1c835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:57Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.390305 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.390366 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.390384 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.390411 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.390431 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:57Z","lastTransitionTime":"2026-02-02T15:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.407885 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da08f9aecc070b31df2ddc1be7c6eb2ad8ef68a931e57ca2e5c5302f4e75751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:57Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.427690 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:57Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.446603 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca57dabf1d1ee5c6fc29291b1be687ca7e97c72650084b02067057fccac41a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:57Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.467505 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:57Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.494635 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.494729 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.494756 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.494788 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.494811 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:57Z","lastTransitionTime":"2026-02-02T15:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.505249 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8206eca4-7d8b-4b59-a4c1-24631dd8cd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41d4a7bdc21f00b5039385d7227d8cb0f4ea3051d3b5065c71a49cb0bfad26a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c11b1859bf3bc0ec4d5e6a58d9becc1ac6333a1ed9820d83d634f1e93221c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c08b087d85be9bb83007c7e38a5de9182a5091716fb8a2ca5ded1b8191bebc40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fd4969f0f16c9aa44208d2bf578f77d9d50914a8cb1764507175aa2eb58e733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59175c8e5a7f3536168a69fd944bf3d536bcd7281eee96002343b7836a54401b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:57Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.533786 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nr6ps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e43cd902-791a-4042-bc08-cf03e075be32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2a89a5d1fc43696f33d313aee5dca316a249fdef6a23ff655e31f14cc449cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07e56782051c93b2a320aa539a42764b97647237d228bed2b6af48cee987c60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07e56782051c93b2a320aa539a42764b97647237d228bed2b6af48cee987c60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce850ac7bd4c2444b65d87232114d55b78d11c6f973537c615900092a45d55e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce850ac7bd4c2444b65d87232114d55b78d11c6f973537c615900092a45d55e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nr6ps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:57Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.569883 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5a81ada-e393-499a-aac6-7b40349958bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4acf2398e2e19bd54f482cb2273c54738e61627187860bb0b6daadddc2565b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831d85055203f153381eb51d98af0cd2e4c4a0309f837545ab3d855a680133e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1380122d1694c1286f1102e6396c6504b91f63c6eb157689dfc0ea7188c17f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5835cb53385359e9c58eaa5d4da2569c313e0fc265ad79a6c124e8cecb1a734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd6dc4bf5b9fd31bfbee6d9340fe981c0c5fca03a79cef9316340ff5a8b3d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db1db75e392ebea8b387b85cabc4276d553bf2c06694cf7f294a383909f32826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58d0beeefc5a719f002da8a57bb8f5209f8e9e527159fcfbf77c28cf1a30fdf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58d0beeefc5a719f002da8a57bb8f5209f8e9e527159fcfbf77c28cf1a30fdf6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T15:15:48Z\\\",\\\"message\\\":\\\" obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI0202 15:15:48.507041 6175 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI0202 15:15:48.507054 6175 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controller-manager-crc in node crc\\\\nI0202 15:15:48.507062 6175 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc after 0 failed attempt(s)\\\\nI0202 15:15:48.507070 6175 default_network_controller.go:776] Recording success event on pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nF0202 15:15:48.507076 6175 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Po\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qxvkf_openshift-ovn-kubernetes(f5a81ada-e393-499a-aac6-7b40349958bb)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efaf2216e4c1750139f080a93b4b608737dab2883de5fddfd9a99248d517b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxvkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:57Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.590014 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8c597" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76cc68c3-9e90-45de-87b2-6ea1b68eadce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ac57831c1c880e64aaa7bedcf999133ee09fc5ecc5124b9d54ecedcc9a13c28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwr7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8c597\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:57Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.600334 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.600399 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.600417 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.600442 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.600462 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:57Z","lastTransitionTime":"2026-02-02T15:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.605147 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hbm26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45553dc5-54ea-4de9-914d-55b8fee477f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d112b1ea41c3d30a7cfa46d17db6aed0999eb0baaaf8a2f21c741d805a21451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d64j4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f76a1f655f871deb7e1364fcc5dbf6b564c4c386300bb49cb75f83518b64286e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d64j4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hbm26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:57Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.624826 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"402a6dc5-6bc7-4d7c-a909-9bb4ca48365d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d5b032497b592363a03f89cb31ba1b551f479004072dddffce540c41190ab81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a3fd02d55a10647428a05a9f53a1429a9827059c523d5deab3c4581c8f1b933\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11600ada13fbc7db1831de385202bc1f041efb656cc58aa49b0f71fe902db9bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fd39f5a05dec7a756dfd519f5df3ef7c77bbb5b8633c5af18a09ff8029cb079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:15:57Z is after 2025-08-24T17:21:41Z" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.704248 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.704717 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.704738 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.704764 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.704783 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:57Z","lastTransitionTime":"2026-02-02T15:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.807678 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.807749 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.807771 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.807800 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.807819 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:57Z","lastTransitionTime":"2026-02-02T15:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.911119 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.911248 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.911272 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.911302 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:57 crc kubenswrapper[4733]: I0202 15:15:57.911323 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:57Z","lastTransitionTime":"2026-02-02T15:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:58 crc kubenswrapper[4733]: I0202 15:15:58.014440 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:58 crc kubenswrapper[4733]: I0202 15:15:58.014500 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:58 crc kubenswrapper[4733]: I0202 15:15:58.014522 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:58 crc kubenswrapper[4733]: I0202 15:15:58.014552 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:58 crc kubenswrapper[4733]: I0202 15:15:58.014573 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:58Z","lastTransitionTime":"2026-02-02T15:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:58 crc kubenswrapper[4733]: I0202 15:15:58.116555 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:58 crc kubenswrapper[4733]: I0202 15:15:58.116600 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:58 crc kubenswrapper[4733]: I0202 15:15:58.116610 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:58 crc kubenswrapper[4733]: I0202 15:15:58.116625 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:58 crc kubenswrapper[4733]: I0202 15:15:58.116637 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:58Z","lastTransitionTime":"2026-02-02T15:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:58 crc kubenswrapper[4733]: I0202 15:15:58.215969 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 19:31:25.067200888 +0000 UTC Feb 02 15:15:58 crc kubenswrapper[4733]: I0202 15:15:58.219191 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:58 crc kubenswrapper[4733]: I0202 15:15:58.219231 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:58 crc kubenswrapper[4733]: I0202 15:15:58.219247 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:58 crc kubenswrapper[4733]: I0202 15:15:58.219265 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:58 crc kubenswrapper[4733]: I0202 15:15:58.219280 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:58Z","lastTransitionTime":"2026-02-02T15:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:58 crc kubenswrapper[4733]: I0202 15:15:58.322518 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:58 crc kubenswrapper[4733]: I0202 15:15:58.322579 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:58 crc kubenswrapper[4733]: I0202 15:15:58.322603 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:58 crc kubenswrapper[4733]: I0202 15:15:58.322634 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:58 crc kubenswrapper[4733]: I0202 15:15:58.322696 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:58Z","lastTransitionTime":"2026-02-02T15:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:58 crc kubenswrapper[4733]: I0202 15:15:58.427156 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:58 crc kubenswrapper[4733]: I0202 15:15:58.427251 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:58 crc kubenswrapper[4733]: I0202 15:15:58.427273 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:58 crc kubenswrapper[4733]: I0202 15:15:58.427303 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:58 crc kubenswrapper[4733]: I0202 15:15:58.427325 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:58Z","lastTransitionTime":"2026-02-02T15:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:58 crc kubenswrapper[4733]: I0202 15:15:58.530238 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:58 crc kubenswrapper[4733]: I0202 15:15:58.530301 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:58 crc kubenswrapper[4733]: I0202 15:15:58.530310 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:58 crc kubenswrapper[4733]: I0202 15:15:58.530357 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:58 crc kubenswrapper[4733]: I0202 15:15:58.530371 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:58Z","lastTransitionTime":"2026-02-02T15:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:58 crc kubenswrapper[4733]: I0202 15:15:58.633481 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:58 crc kubenswrapper[4733]: I0202 15:15:58.633544 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:58 crc kubenswrapper[4733]: I0202 15:15:58.633568 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:58 crc kubenswrapper[4733]: I0202 15:15:58.633600 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:58 crc kubenswrapper[4733]: I0202 15:15:58.633623 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:58Z","lastTransitionTime":"2026-02-02T15:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:58 crc kubenswrapper[4733]: I0202 15:15:58.736999 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:58 crc kubenswrapper[4733]: I0202 15:15:58.737090 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:58 crc kubenswrapper[4733]: I0202 15:15:58.737116 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:58 crc kubenswrapper[4733]: I0202 15:15:58.737148 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:58 crc kubenswrapper[4733]: I0202 15:15:58.737214 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:58Z","lastTransitionTime":"2026-02-02T15:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:58 crc kubenswrapper[4733]: I0202 15:15:58.840146 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:58 crc kubenswrapper[4733]: I0202 15:15:58.840301 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:58 crc kubenswrapper[4733]: I0202 15:15:58.840326 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:58 crc kubenswrapper[4733]: I0202 15:15:58.840351 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:58 crc kubenswrapper[4733]: I0202 15:15:58.840368 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:58Z","lastTransitionTime":"2026-02-02T15:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:58 crc kubenswrapper[4733]: I0202 15:15:58.943130 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:58 crc kubenswrapper[4733]: I0202 15:15:58.943225 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:58 crc kubenswrapper[4733]: I0202 15:15:58.943270 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:58 crc kubenswrapper[4733]: I0202 15:15:58.943311 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:58 crc kubenswrapper[4733]: I0202 15:15:58.943335 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:58Z","lastTransitionTime":"2026-02-02T15:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.046492 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.046567 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.046579 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.046596 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.046609 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:59Z","lastTransitionTime":"2026-02-02T15:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.149456 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.149512 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.149526 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.149545 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.149562 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:59Z","lastTransitionTime":"2026-02-02T15:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.216863 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-14 04:16:50.060145215 +0000 UTC Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.252653 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.252734 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.252754 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.252777 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.252796 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:59Z","lastTransitionTime":"2026-02-02T15:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.255066 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.255156 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.255205 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:15:59 crc kubenswrapper[4733]: E0202 15:15:59.255273 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:15:59 crc kubenswrapper[4733]: E0202 15:15:59.255416 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.255442 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:15:59 crc kubenswrapper[4733]: E0202 15:15:59.255587 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:15:59 crc kubenswrapper[4733]: E0202 15:15:59.255787 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.355610 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.355666 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.355691 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.355720 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.355742 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:59Z","lastTransitionTime":"2026-02-02T15:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.459007 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.459064 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.459081 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.459104 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.459121 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:59Z","lastTransitionTime":"2026-02-02T15:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.563707 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.563778 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.563794 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.563829 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.563859 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:59Z","lastTransitionTime":"2026-02-02T15:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.666957 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.667022 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.667039 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.667065 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.667085 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:59Z","lastTransitionTime":"2026-02-02T15:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.770642 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.770724 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.770744 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.770770 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.770789 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:59Z","lastTransitionTime":"2026-02-02T15:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.874360 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.874414 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.874429 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.874447 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.874458 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:59Z","lastTransitionTime":"2026-02-02T15:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.977893 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.977966 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.977984 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.978013 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:15:59 crc kubenswrapper[4733]: I0202 15:15:59.978035 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:15:59Z","lastTransitionTime":"2026-02-02T15:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:00 crc kubenswrapper[4733]: I0202 15:16:00.080582 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:00 crc kubenswrapper[4733]: I0202 15:16:00.080642 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:00 crc kubenswrapper[4733]: I0202 15:16:00.080691 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:00 crc kubenswrapper[4733]: I0202 15:16:00.080714 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:00 crc kubenswrapper[4733]: I0202 15:16:00.080725 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:00Z","lastTransitionTime":"2026-02-02T15:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:00 crc kubenswrapper[4733]: I0202 15:16:00.183922 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:00 crc kubenswrapper[4733]: I0202 15:16:00.184009 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:00 crc kubenswrapper[4733]: I0202 15:16:00.184030 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:00 crc kubenswrapper[4733]: I0202 15:16:00.184057 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:00 crc kubenswrapper[4733]: I0202 15:16:00.184076 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:00Z","lastTransitionTime":"2026-02-02T15:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:00 crc kubenswrapper[4733]: I0202 15:16:00.217749 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-30 00:38:52.096257012 +0000 UTC Feb 02 15:16:00 crc kubenswrapper[4733]: I0202 15:16:00.286843 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:00 crc kubenswrapper[4733]: I0202 15:16:00.286928 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:00 crc kubenswrapper[4733]: I0202 15:16:00.286953 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:00 crc kubenswrapper[4733]: I0202 15:16:00.286982 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:00 crc kubenswrapper[4733]: I0202 15:16:00.287008 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:00Z","lastTransitionTime":"2026-02-02T15:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:00 crc kubenswrapper[4733]: I0202 15:16:00.389580 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:00 crc kubenswrapper[4733]: I0202 15:16:00.389647 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:00 crc kubenswrapper[4733]: I0202 15:16:00.389666 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:00 crc kubenswrapper[4733]: I0202 15:16:00.389691 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:00 crc kubenswrapper[4733]: I0202 15:16:00.389711 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:00Z","lastTransitionTime":"2026-02-02T15:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:00 crc kubenswrapper[4733]: I0202 15:16:00.492561 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:00 crc kubenswrapper[4733]: I0202 15:16:00.492641 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:00 crc kubenswrapper[4733]: I0202 15:16:00.492666 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:00 crc kubenswrapper[4733]: I0202 15:16:00.492698 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:00 crc kubenswrapper[4733]: I0202 15:16:00.492720 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:00Z","lastTransitionTime":"2026-02-02T15:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:00 crc kubenswrapper[4733]: I0202 15:16:00.596134 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:00 crc kubenswrapper[4733]: I0202 15:16:00.596237 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:00 crc kubenswrapper[4733]: I0202 15:16:00.596260 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:00 crc kubenswrapper[4733]: I0202 15:16:00.596283 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:00 crc kubenswrapper[4733]: I0202 15:16:00.596300 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:00Z","lastTransitionTime":"2026-02-02T15:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:00 crc kubenswrapper[4733]: I0202 15:16:00.699744 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:00 crc kubenswrapper[4733]: I0202 15:16:00.699811 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:00 crc kubenswrapper[4733]: I0202 15:16:00.699828 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:00 crc kubenswrapper[4733]: I0202 15:16:00.699852 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:00 crc kubenswrapper[4733]: I0202 15:16:00.699869 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:00Z","lastTransitionTime":"2026-02-02T15:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:00 crc kubenswrapper[4733]: I0202 15:16:00.802649 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:00 crc kubenswrapper[4733]: I0202 15:16:00.802724 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:00 crc kubenswrapper[4733]: I0202 15:16:00.802746 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:00 crc kubenswrapper[4733]: I0202 15:16:00.802778 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:00 crc kubenswrapper[4733]: I0202 15:16:00.802802 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:00Z","lastTransitionTime":"2026-02-02T15:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:00 crc kubenswrapper[4733]: I0202 15:16:00.905646 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:00 crc kubenswrapper[4733]: I0202 15:16:00.905704 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:00 crc kubenswrapper[4733]: I0202 15:16:00.905720 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:00 crc kubenswrapper[4733]: I0202 15:16:00.905741 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:00 crc kubenswrapper[4733]: I0202 15:16:00.905759 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:00Z","lastTransitionTime":"2026-02-02T15:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.008785 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.008854 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.008876 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.008905 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.008928 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:01Z","lastTransitionTime":"2026-02-02T15:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.111545 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.111603 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.111620 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.111643 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.111663 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:01Z","lastTransitionTime":"2026-02-02T15:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.215074 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.215134 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.215151 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.215227 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.215253 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:01Z","lastTransitionTime":"2026-02-02T15:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.218390 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 22:52:22.555950537 +0000 UTC Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.255092 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.255218 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.255329 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:16:01 crc kubenswrapper[4733]: E0202 15:16:01.255329 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.255148 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:16:01 crc kubenswrapper[4733]: E0202 15:16:01.255565 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:16:01 crc kubenswrapper[4733]: E0202 15:16:01.256372 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:16:01 crc kubenswrapper[4733]: E0202 15:16:01.256551 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.256668 4733 scope.go:117] "RemoveContainer" containerID="58d0beeefc5a719f002da8a57bb8f5209f8e9e527159fcfbf77c28cf1a30fdf6" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.319145 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.319256 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.319280 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.319311 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.319340 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:01Z","lastTransitionTime":"2026-02-02T15:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.422114 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.422208 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.422232 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.422260 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.422284 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:01Z","lastTransitionTime":"2026-02-02T15:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.526673 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.526741 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.526765 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.526798 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.526820 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:01Z","lastTransitionTime":"2026-02-02T15:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.591712 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxvkf_f5a81ada-e393-499a-aac6-7b40349958bb/ovnkube-controller/1.log" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.596727 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" event={"ID":"f5a81ada-e393-499a-aac6-7b40349958bb","Type":"ContainerStarted","Data":"fa356e9d1cf1236a6d325e762401b42848ed4c48bc4c538a5844e09de54d16c1"} Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.597357 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.619227 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"402a6dc5-6bc7-4d7c-a909-9bb4ca48365d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d5b032497b592363a03f89cb31ba1b551f479004072dddffce540c41190ab81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a3fd02d55a10647428a05a9f53a1429a9827059c523d5deab3c4581c8f1b933\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11600ada13fbc7db1831de385202bc1f041efb656cc58aa49b0f71fe902db9bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fd39f5a05dec7a756dfd519f5df3ef7c77bbb5b8633c5af18a09ff8029cb079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:01Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.635840 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.635881 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.635899 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.636088 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.636108 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:01Z","lastTransitionTime":"2026-02-02T15:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.651138 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b38b8410bea95465ee58803c4160039bf7c082ca47893680da7469418f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27003b336112760daa2079f4819641850c27fbc8415b63b5094c86195291aca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:01Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.672686 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jh6vk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76a05d6-d9c7-4183-ad90-b3131f225841\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f2659e2a914d3191daddbfb5ad109e487f03dabb559d324244e127add7d41d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jh6vk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:01Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.688887 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83d0e810-ea5c-47aa-947b-7865b045bf94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6144eb32f88bc106f7724e48133b568a137c4f3fbbfe30c7fcef405d3ab735a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51318ad71b42f4cc8d376aa50e44580e68da89d83d80defb996329a2ee5c2ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zgsv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:01Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.706905 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:01Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.721534 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da08f9aecc070b31df2ddc1be7c6eb2ad8ef68a931e57ca2e5c5302f4e75751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:01Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.735273 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:01Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.740258 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.740313 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.740324 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.740341 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.740352 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:01Z","lastTransitionTime":"2026-02-02T15:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.753609 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca57dabf1d1ee5c6fc29291b1be687ca7e97c72650084b02067057fccac41a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:01Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.770314 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:01Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.786081 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5xgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56681968-f264-440e-aa12-d4f81879efe9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e250891b1065a38a448aa6abd300b37543e819e87f006b03cc37983cb643982a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdjjl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5xgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:01Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.800890 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7r9fd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4dfd44b-b363-4162-9f43-9a8dc4a3b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7r9fd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:01Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.824351 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7082da2-f2e7-4b73-ba35-504e5cd4c920\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd383d69b9b27d52fa8672ff6e0294a74876d676973bdc0ca95125fef5f2d085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037e379a109ce230098e3ca223fd1dea3149de3ebe601f56e2bb295f3093a3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6fdd19dac506f2d5d7eedc44ec2cda92a9381b73e91a45d949e4fcde4dfc283\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13965f5abcfb0a36c33310518955f192673e630a1aaba94b49400026de5bc2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://992f4198192cbe4b1c89d575aa4a20c29c1115fd5eaa8986c6ed70c1d6b143c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0202 15:15:31.010033 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 15:15:31.011243 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3543528698/tls.crt::/tmp/serving-cert-3543528698/tls.key\\\\\\\"\\\\nI0202 15:15:36.416658 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 15:15:36.428087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 15:15:36.428134 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 15:15:36.428243 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 15:15:36.428280 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 15:15:36.439678 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 15:15:36.439710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439717 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439722 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 15:15:36.439726 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 15:15:36.439729 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 15:15:36.439732 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 15:15:36.439753 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 15:15:36.450716 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://226f5949820d4f498dbf6fa251c6edf1c6356900bd946e6afae2c29042b1c835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:01Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.842957 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.843000 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.843013 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.843031 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.843047 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:01Z","lastTransitionTime":"2026-02-02T15:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.844106 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nr6ps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e43cd902-791a-4042-bc08-cf03e075be32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2a89a5d1fc43696f33d313aee5dca316a249fdef6a23ff655e31f14cc449cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07e56782051c93b2a320aa539a42764b97647237d228bed2b6af48cee987c60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07e56782051c93b2a320aa539a42764b97647237d228bed2b6af48cee987c60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce850ac7bd4c2444b65d87232114d55b78d11c6f973537c615900092a45d55e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce850ac7bd4c2444b65d87232114d55b78d11c6f973537c615900092a45d55e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nr6ps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:01Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.871417 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5a81ada-e393-499a-aac6-7b40349958bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4acf2398e2e19bd54f482cb2273c54738e61627187860bb0b6daadddc2565b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831d85055203f153381eb51d98af0cd2e4c4a0309f837545ab3d855a680133e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1380122d1694c1286f1102e6396c6504b91f63c6eb157689dfc0ea7188c17f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5835cb53385359e9c58eaa5d4da2569c313e0fc265ad79a6c124e8cecb1a734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd6dc4bf5b9fd31bfbee6d9340fe981c0c5fca03a79cef9316340ff5a8b3d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db1db75e392ebea8b387b85cabc4276d553bf2c06694cf7f294a383909f32826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa356e9d1cf1236a6d325e762401b42848ed4c48bc4c538a5844e09de54d16c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58d0beeefc5a719f002da8a57bb8f5209f8e9e527159fcfbf77c28cf1a30fdf6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T15:15:48Z\\\",\\\"message\\\":\\\" obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI0202 15:15:48.507041 6175 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI0202 15:15:48.507054 6175 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controller-manager-crc in node crc\\\\nI0202 15:15:48.507062 6175 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc after 0 failed attempt(s)\\\\nI0202 15:15:48.507070 6175 default_network_controller.go:776] Recording success event on pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nF0202 15:15:48.507076 6175 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Po\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:16:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efaf2216e4c1750139f080a93b4b608737dab2883de5fddfd9a99248d517b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxvkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:01Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.898550 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8c597" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76cc68c3-9e90-45de-87b2-6ea1b68eadce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ac57831c1c880e64aaa7bedcf999133ee09fc5ecc5124b9d54ecedcc9a13c28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwr7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8c597\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:01Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.910578 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hbm26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45553dc5-54ea-4de9-914d-55b8fee477f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d112b1ea41c3d30a7cfa46d17db6aed0999eb0baaaf8a2f21c741d805a21451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d64j4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f76a1f655f871deb7e1364fcc5dbf6b564c4c386300bb49cb75f83518b64286e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d64j4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hbm26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:01Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.931734 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8206eca4-7d8b-4b59-a4c1-24631dd8cd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41d4a7bdc21f00b5039385d7227d8cb0f4ea3051d3b5065c71a49cb0bfad26a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c11b1859bf3bc0ec4d5e6a58d9becc1ac6333a1ed9820d83d634f1e93221c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c08b087d85be9bb83007c7e38a5de9182a5091716fb8a2ca5ded1b8191bebc40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fd4969f0f16c9aa44208d2bf578f77d9d50914a8cb1764507175aa2eb58e733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59175c8e5a7f3536168a69fd944bf3d536bcd7281eee96002343b7836a54401b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:01Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.945182 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.945221 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.945231 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.945249 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:01 crc kubenswrapper[4733]: I0202 15:16:01.945261 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:01Z","lastTransitionTime":"2026-02-02T15:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.048127 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.048195 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.048210 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.048232 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.048244 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:02Z","lastTransitionTime":"2026-02-02T15:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.150711 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.150757 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.150766 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.150781 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.150792 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:02Z","lastTransitionTime":"2026-02-02T15:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.218704 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-14 02:59:32.038178336 +0000 UTC Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.254484 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.254567 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.254588 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.254623 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.254646 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:02Z","lastTransitionTime":"2026-02-02T15:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.357459 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.357501 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.357513 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.357531 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.357543 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:02Z","lastTransitionTime":"2026-02-02T15:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.460695 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.460757 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.460775 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.460799 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.460817 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:02Z","lastTransitionTime":"2026-02-02T15:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.564567 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.564623 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.564639 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.564658 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.564706 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:02Z","lastTransitionTime":"2026-02-02T15:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.602586 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxvkf_f5a81ada-e393-499a-aac6-7b40349958bb/ovnkube-controller/2.log" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.603633 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxvkf_f5a81ada-e393-499a-aac6-7b40349958bb/ovnkube-controller/1.log" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.607704 4733 generic.go:334] "Generic (PLEG): container finished" podID="f5a81ada-e393-499a-aac6-7b40349958bb" containerID="fa356e9d1cf1236a6d325e762401b42848ed4c48bc4c538a5844e09de54d16c1" exitCode=1 Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.607751 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" event={"ID":"f5a81ada-e393-499a-aac6-7b40349958bb","Type":"ContainerDied","Data":"fa356e9d1cf1236a6d325e762401b42848ed4c48bc4c538a5844e09de54d16c1"} Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.607798 4733 scope.go:117] "RemoveContainer" containerID="58d0beeefc5a719f002da8a57bb8f5209f8e9e527159fcfbf77c28cf1a30fdf6" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.609098 4733 scope.go:117] "RemoveContainer" containerID="fa356e9d1cf1236a6d325e762401b42848ed4c48bc4c538a5844e09de54d16c1" Feb 02 15:16:02 crc kubenswrapper[4733]: E0202 15:16:02.609557 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qxvkf_openshift-ovn-kubernetes(f5a81ada-e393-499a-aac6-7b40349958bb)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.631313 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:02Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.652149 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b38b8410bea95465ee58803c4160039bf7c082ca47893680da7469418f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27003b336112760daa2079f4819641850c27fbc8415b63b5094c86195291aca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:02Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.667234 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.667311 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.667330 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.667356 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.667375 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:02Z","lastTransitionTime":"2026-02-02T15:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.672739 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jh6vk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76a05d6-d9c7-4183-ad90-b3131f225841\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f2659e2a914d3191daddbfb5ad109e487f03dabb559d324244e127add7d41d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jh6vk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:02Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.692555 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83d0e810-ea5c-47aa-947b-7865b045bf94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6144eb32f88bc106f7724e48133b568a137c4f3fbbfe30c7fcef405d3ab735a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51318ad71b42f4cc8d376aa50e44580e68da89d83d80defb996329a2ee5c2ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zgsv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:02Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.711628 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca57dabf1d1ee5c6fc29291b1be687ca7e97c72650084b02067057fccac41a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:02Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.731366 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:02Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.747384 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5xgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56681968-f264-440e-aa12-d4f81879efe9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e250891b1065a38a448aa6abd300b37543e819e87f006b03cc37983cb643982a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdjjl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5xgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:02Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.763751 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7r9fd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4dfd44b-b363-4162-9f43-9a8dc4a3b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7r9fd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:02Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.773745 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.773799 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.773812 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.773830 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.773848 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:02Z","lastTransitionTime":"2026-02-02T15:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.791528 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7082da2-f2e7-4b73-ba35-504e5cd4c920\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd383d69b9b27d52fa8672ff6e0294a74876d676973bdc0ca95125fef5f2d085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037e379a109ce230098e3ca223fd1dea3149de3ebe601f56e2bb295f3093a3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6fdd19dac506f2d5d7eedc44ec2cda92a9381b73e91a45d949e4fcde4dfc283\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13965f5abcfb0a36c33310518955f192673e630a1aaba94b49400026de5bc2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://992f4198192cbe4b1c89d575aa4a20c29c1115fd5eaa8986c6ed70c1d6b143c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0202 15:15:31.010033 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 15:15:31.011243 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3543528698/tls.crt::/tmp/serving-cert-3543528698/tls.key\\\\\\\"\\\\nI0202 15:15:36.416658 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 15:15:36.428087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 15:15:36.428134 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 15:15:36.428243 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 15:15:36.428280 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 15:15:36.439678 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 15:15:36.439710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439717 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439722 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 15:15:36.439726 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 15:15:36.439729 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 15:15:36.439732 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 15:15:36.439753 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 15:15:36.450716 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://226f5949820d4f498dbf6fa251c6edf1c6356900bd946e6afae2c29042b1c835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:02Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.813266 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da08f9aecc070b31df2ddc1be7c6eb2ad8ef68a931e57ca2e5c5302f4e75751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:02Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.834238 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:02Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.853009 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hbm26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45553dc5-54ea-4de9-914d-55b8fee477f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d112b1ea41c3d30a7cfa46d17db6aed0999eb0baaaf8a2f21c741d805a21451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d64j4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f76a1f655f871deb7e1364fcc5dbf6b564c4c386300bb49cb75f83518b64286e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d64j4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hbm26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:02Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.876893 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.876987 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.877022 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.877050 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.877074 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:02Z","lastTransitionTime":"2026-02-02T15:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.890001 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8206eca4-7d8b-4b59-a4c1-24631dd8cd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41d4a7bdc21f00b5039385d7227d8cb0f4ea3051d3b5065c71a49cb0bfad26a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c11b1859bf3bc0ec4d5e6a58d9becc1ac6333a1ed9820d83d634f1e93221c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c08b087d85be9bb83007c7e38a5de9182a5091716fb8a2ca5ded1b8191bebc40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fd4969f0f16c9aa44208d2bf578f77d9d50914a8cb1764507175aa2eb58e733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59175c8e5a7f3536168a69fd944bf3d536bcd7281eee96002343b7836a54401b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:02Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.913288 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nr6ps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e43cd902-791a-4042-bc08-cf03e075be32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2a89a5d1fc43696f33d313aee5dca316a249fdef6a23ff655e31f14cc449cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07e56782051c93b2a320aa539a42764b97647237d228bed2b6af48cee987c60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07e56782051c93b2a320aa539a42764b97647237d228bed2b6af48cee987c60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce850ac7bd4c2444b65d87232114d55b78d11c6f973537c615900092a45d55e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce850ac7bd4c2444b65d87232114d55b78d11c6f973537c615900092a45d55e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nr6ps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:02Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.945685 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5a81ada-e393-499a-aac6-7b40349958bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4acf2398e2e19bd54f482cb2273c54738e61627187860bb0b6daadddc2565b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831d85055203f153381eb51d98af0cd2e4c4a0309f837545ab3d855a680133e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1380122d1694c1286f1102e6396c6504b91f63c6eb157689dfc0ea7188c17f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5835cb53385359e9c58eaa5d4da2569c313e0fc265ad79a6c124e8cecb1a734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd6dc4bf5b9fd31bfbee6d9340fe981c0c5fca03a79cef9316340ff5a8b3d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db1db75e392ebea8b387b85cabc4276d553bf2c06694cf7f294a383909f32826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa356e9d1cf1236a6d325e762401b42848ed4c48bc4c538a5844e09de54d16c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58d0beeefc5a719f002da8a57bb8f5209f8e9e527159fcfbf77c28cf1a30fdf6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T15:15:48Z\\\",\\\"message\\\":\\\" obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI0202 15:15:48.507041 6175 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI0202 15:15:48.507054 6175 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controller-manager-crc in node crc\\\\nI0202 15:15:48.507062 6175 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc after 0 failed attempt(s)\\\\nI0202 15:15:48.507070 6175 default_network_controller.go:776] Recording success event on pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nF0202 15:15:48.507076 6175 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Po\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa356e9d1cf1236a6d325e762401b42848ed4c48bc4c538a5844e09de54d16c1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T15:16:02Z\\\",\\\"message\\\":\\\"\\\\nI0202 15:16:02.301761 6382 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-5xgpn in node crc\\\\nI0202 15:16:02.301771 6382 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-8c597\\\\nI0202 15:16:02.301779 6382 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-5xgpn after 0 failed attempt(s)\\\\nI0202 15:16:02.301784 6382 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-8c597 in node crc\\\\nI0202 15:16:02.301793 6382 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-5xgpn\\\\nI0202 15:16:02.301794 6382 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-8c597 after 0 failed attempt(s)\\\\nI0202 15:16:02.301805 6382 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-8c597\\\\nI0202 15:16:02.301355 6382 services_controller.go:360] Finished syncing service metrics on namespace openshift-ingress-operator for network=default : 1.801637ms\\\\nI0202 15:16:02.301684 6382 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI0202 15:16:02.301833 6382 services_controller.go:356] Processing sync for service openshift-console-operator/metrics for network=default\\\\nI0202 15:16:02\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:16:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efaf2216e4c1750139f080a93b4b608737dab2883de5fddfd9a99248d517b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxvkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:02Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.962664 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8c597" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76cc68c3-9e90-45de-87b2-6ea1b68eadce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ac57831c1c880e64aaa7bedcf999133ee09fc5ecc5124b9d54ecedcc9a13c28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwr7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8c597\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:02Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.980567 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.980627 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.980644 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.980666 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.980683 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:02Z","lastTransitionTime":"2026-02-02T15:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:02 crc kubenswrapper[4733]: I0202 15:16:02.983322 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"402a6dc5-6bc7-4d7c-a909-9bb4ca48365d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d5b032497b592363a03f89cb31ba1b551f479004072dddffce540c41190ab81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a3fd02d55a10647428a05a9f53a1429a9827059c523d5deab3c4581c8f1b933\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11600ada13fbc7db1831de385202bc1f041efb656cc58aa49b0f71fe902db9bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fd39f5a05dec7a756dfd519f5df3ef7c77bbb5b8633c5af18a09ff8029cb079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:02Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.083277 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.083335 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.083352 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.083373 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.083390 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:03Z","lastTransitionTime":"2026-02-02T15:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.186297 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.186345 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.186362 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.186383 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.186401 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:03Z","lastTransitionTime":"2026-02-02T15:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.219228 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 10:07:41.918697584 +0000 UTC Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.255136 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.255144 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.255269 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:16:03 crc kubenswrapper[4733]: E0202 15:16:03.255439 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.255515 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:16:03 crc kubenswrapper[4733]: E0202 15:16:03.255661 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:16:03 crc kubenswrapper[4733]: E0202 15:16:03.255805 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:16:03 crc kubenswrapper[4733]: E0202 15:16:03.255917 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.288478 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.288532 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.288552 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.288579 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.288600 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:03Z","lastTransitionTime":"2026-02-02T15:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.392550 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.392605 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.392629 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.392660 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.392683 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:03Z","lastTransitionTime":"2026-02-02T15:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.496235 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.496284 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.496313 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.496339 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.496358 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:03Z","lastTransitionTime":"2026-02-02T15:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.599744 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.599822 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.599846 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.599872 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.599889 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:03Z","lastTransitionTime":"2026-02-02T15:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.612852 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxvkf_f5a81ada-e393-499a-aac6-7b40349958bb/ovnkube-controller/2.log" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.618139 4733 scope.go:117] "RemoveContainer" containerID="fa356e9d1cf1236a6d325e762401b42848ed4c48bc4c538a5844e09de54d16c1" Feb 02 15:16:03 crc kubenswrapper[4733]: E0202 15:16:03.618567 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qxvkf_openshift-ovn-kubernetes(f5a81ada-e393-499a-aac6-7b40349958bb)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.636012 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da08f9aecc070b31df2ddc1be7c6eb2ad8ef68a931e57ca2e5c5302f4e75751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:03Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.652296 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:03Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.666503 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca57dabf1d1ee5c6fc29291b1be687ca7e97c72650084b02067057fccac41a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:03Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.691778 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:03Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.702677 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.702733 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.702753 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.702776 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.702792 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:03Z","lastTransitionTime":"2026-02-02T15:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.707292 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5xgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56681968-f264-440e-aa12-d4f81879efe9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e250891b1065a38a448aa6abd300b37543e819e87f006b03cc37983cb643982a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdjjl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5xgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:03Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.721208 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7r9fd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4dfd44b-b363-4162-9f43-9a8dc4a3b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7r9fd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:03Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.741194 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7082da2-f2e7-4b73-ba35-504e5cd4c920\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd383d69b9b27d52fa8672ff6e0294a74876d676973bdc0ca95125fef5f2d085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037e379a109ce230098e3ca223fd1dea3149de3ebe601f56e2bb295f3093a3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6fdd19dac506f2d5d7eedc44ec2cda92a9381b73e91a45d949e4fcde4dfc283\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13965f5abcfb0a36c33310518955f192673e630a1aaba94b49400026de5bc2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://992f4198192cbe4b1c89d575aa4a20c29c1115fd5eaa8986c6ed70c1d6b143c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0202 15:15:31.010033 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 15:15:31.011243 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3543528698/tls.crt::/tmp/serving-cert-3543528698/tls.key\\\\\\\"\\\\nI0202 15:15:36.416658 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 15:15:36.428087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 15:15:36.428134 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 15:15:36.428243 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 15:15:36.428280 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 15:15:36.439678 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 15:15:36.439710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439717 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439722 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 15:15:36.439726 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 15:15:36.439729 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 15:15:36.439732 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 15:15:36.439753 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 15:15:36.450716 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://226f5949820d4f498dbf6fa251c6edf1c6356900bd946e6afae2c29042b1c835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:03Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.769277 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5a81ada-e393-499a-aac6-7b40349958bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4acf2398e2e19bd54f482cb2273c54738e61627187860bb0b6daadddc2565b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831d85055203f153381eb51d98af0cd2e4c4a0309f837545ab3d855a680133e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1380122d1694c1286f1102e6396c6504b91f63c6eb157689dfc0ea7188c17f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5835cb53385359e9c58eaa5d4da2569c313e0fc265ad79a6c124e8cecb1a734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd6dc4bf5b9fd31bfbee6d9340fe981c0c5fca03a79cef9316340ff5a8b3d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db1db75e392ebea8b387b85cabc4276d553bf2c06694cf7f294a383909f32826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa356e9d1cf1236a6d325e762401b42848ed4c48bc4c538a5844e09de54d16c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa356e9d1cf1236a6d325e762401b42848ed4c48bc4c538a5844e09de54d16c1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T15:16:02Z\\\",\\\"message\\\":\\\"\\\\nI0202 15:16:02.301761 6382 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-5xgpn in node crc\\\\nI0202 15:16:02.301771 6382 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-8c597\\\\nI0202 15:16:02.301779 6382 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-5xgpn after 0 failed attempt(s)\\\\nI0202 15:16:02.301784 6382 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-8c597 in node crc\\\\nI0202 15:16:02.301793 6382 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-5xgpn\\\\nI0202 15:16:02.301794 6382 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-8c597 after 0 failed attempt(s)\\\\nI0202 15:16:02.301805 6382 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-8c597\\\\nI0202 15:16:02.301355 6382 services_controller.go:360] Finished syncing service metrics on namespace openshift-ingress-operator for network=default : 1.801637ms\\\\nI0202 15:16:02.301684 6382 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI0202 15:16:02.301833 6382 services_controller.go:356] Processing sync for service openshift-console-operator/metrics for network=default\\\\nI0202 15:16:02\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:16:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qxvkf_openshift-ovn-kubernetes(f5a81ada-e393-499a-aac6-7b40349958bb)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efaf2216e4c1750139f080a93b4b608737dab2883de5fddfd9a99248d517b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxvkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:03Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.784249 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8c597" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76cc68c3-9e90-45de-87b2-6ea1b68eadce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ac57831c1c880e64aaa7bedcf999133ee09fc5ecc5124b9d54ecedcc9a13c28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwr7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8c597\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:03Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.801344 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hbm26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45553dc5-54ea-4de9-914d-55b8fee477f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d112b1ea41c3d30a7cfa46d17db6aed0999eb0baaaf8a2f21c741d805a21451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d64j4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f76a1f655f871deb7e1364fcc5dbf6b564c4c386300bb49cb75f83518b64286e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d64j4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hbm26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:03Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.806092 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.806209 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.806237 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.806283 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.806307 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:03Z","lastTransitionTime":"2026-02-02T15:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.839329 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8206eca4-7d8b-4b59-a4c1-24631dd8cd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41d4a7bdc21f00b5039385d7227d8cb0f4ea3051d3b5065c71a49cb0bfad26a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c11b1859bf3bc0ec4d5e6a58d9becc1ac6333a1ed9820d83d634f1e93221c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c08b087d85be9bb83007c7e38a5de9182a5091716fb8a2ca5ded1b8191bebc40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fd4969f0f16c9aa44208d2bf578f77d9d50914a8cb1764507175aa2eb58e733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59175c8e5a7f3536168a69fd944bf3d536bcd7281eee96002343b7836a54401b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:03Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.863998 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nr6ps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e43cd902-791a-4042-bc08-cf03e075be32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2a89a5d1fc43696f33d313aee5dca316a249fdef6a23ff655e31f14cc449cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07e56782051c93b2a320aa539a42764b97647237d228bed2b6af48cee987c60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07e56782051c93b2a320aa539a42764b97647237d228bed2b6af48cee987c60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce850ac7bd4c2444b65d87232114d55b78d11c6f973537c615900092a45d55e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce850ac7bd4c2444b65d87232114d55b78d11c6f973537c615900092a45d55e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nr6ps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:03Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.881798 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"402a6dc5-6bc7-4d7c-a909-9bb4ca48365d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d5b032497b592363a03f89cb31ba1b551f479004072dddffce540c41190ab81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a3fd02d55a10647428a05a9f53a1429a9827059c523d5deab3c4581c8f1b933\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11600ada13fbc7db1831de385202bc1f041efb656cc58aa49b0f71fe902db9bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fd39f5a05dec7a756dfd519f5df3ef7c77bbb5b8633c5af18a09ff8029cb079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:03Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.902497 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jh6vk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76a05d6-d9c7-4183-ad90-b3131f225841\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f2659e2a914d3191daddbfb5ad109e487f03dabb559d324244e127add7d41d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jh6vk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:03Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.909484 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.909547 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.909570 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.909603 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.909624 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:03Z","lastTransitionTime":"2026-02-02T15:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.924910 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83d0e810-ea5c-47aa-947b-7865b045bf94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6144eb32f88bc106f7724e48133b568a137c4f3fbbfe30c7fcef405d3ab735a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51318ad71b42f4cc8d376aa50e44580e68da89d83d80defb996329a2ee5c2ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zgsv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:03Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.944225 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:03Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:03 crc kubenswrapper[4733]: I0202 15:16:03.964871 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b38b8410bea95465ee58803c4160039bf7c082ca47893680da7469418f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27003b336112760daa2079f4819641850c27fbc8415b63b5094c86195291aca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:03Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:04 crc kubenswrapper[4733]: I0202 15:16:04.013143 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:04 crc kubenswrapper[4733]: I0202 15:16:04.013238 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:04 crc kubenswrapper[4733]: I0202 15:16:04.013256 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:04 crc kubenswrapper[4733]: I0202 15:16:04.013278 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:04 crc kubenswrapper[4733]: I0202 15:16:04.013295 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:04Z","lastTransitionTime":"2026-02-02T15:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:04 crc kubenswrapper[4733]: I0202 15:16:04.116124 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:04 crc kubenswrapper[4733]: I0202 15:16:04.116220 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:04 crc kubenswrapper[4733]: I0202 15:16:04.116241 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:04 crc kubenswrapper[4733]: I0202 15:16:04.116263 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:04 crc kubenswrapper[4733]: I0202 15:16:04.116281 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:04Z","lastTransitionTime":"2026-02-02T15:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:04 crc kubenswrapper[4733]: I0202 15:16:04.219546 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:04 crc kubenswrapper[4733]: I0202 15:16:04.219615 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:04 crc kubenswrapper[4733]: I0202 15:16:04.219633 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:04 crc kubenswrapper[4733]: I0202 15:16:04.219658 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:04 crc kubenswrapper[4733]: I0202 15:16:04.219677 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:04Z","lastTransitionTime":"2026-02-02T15:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:04 crc kubenswrapper[4733]: I0202 15:16:04.219354 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-16 12:11:56.87974081 +0000 UTC Feb 02 15:16:04 crc kubenswrapper[4733]: I0202 15:16:04.322589 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:04 crc kubenswrapper[4733]: I0202 15:16:04.322665 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:04 crc kubenswrapper[4733]: I0202 15:16:04.322695 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:04 crc kubenswrapper[4733]: I0202 15:16:04.322715 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:04 crc kubenswrapper[4733]: I0202 15:16:04.322729 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:04Z","lastTransitionTime":"2026-02-02T15:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:04 crc kubenswrapper[4733]: I0202 15:16:04.425955 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:04 crc kubenswrapper[4733]: I0202 15:16:04.426024 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:04 crc kubenswrapper[4733]: I0202 15:16:04.426042 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:04 crc kubenswrapper[4733]: I0202 15:16:04.426066 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:04 crc kubenswrapper[4733]: I0202 15:16:04.426092 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:04Z","lastTransitionTime":"2026-02-02T15:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:04 crc kubenswrapper[4733]: I0202 15:16:04.529491 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:04 crc kubenswrapper[4733]: I0202 15:16:04.529547 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:04 crc kubenswrapper[4733]: I0202 15:16:04.529574 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:04 crc kubenswrapper[4733]: I0202 15:16:04.529604 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:04 crc kubenswrapper[4733]: I0202 15:16:04.529629 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:04Z","lastTransitionTime":"2026-02-02T15:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:04 crc kubenswrapper[4733]: I0202 15:16:04.632200 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:04 crc kubenswrapper[4733]: I0202 15:16:04.632278 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:04 crc kubenswrapper[4733]: I0202 15:16:04.632301 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:04 crc kubenswrapper[4733]: I0202 15:16:04.632330 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:04 crc kubenswrapper[4733]: I0202 15:16:04.632352 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:04Z","lastTransitionTime":"2026-02-02T15:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:04 crc kubenswrapper[4733]: I0202 15:16:04.734947 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:04 crc kubenswrapper[4733]: I0202 15:16:04.735049 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:04 crc kubenswrapper[4733]: I0202 15:16:04.735074 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:04 crc kubenswrapper[4733]: I0202 15:16:04.735098 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:04 crc kubenswrapper[4733]: I0202 15:16:04.735155 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:04Z","lastTransitionTime":"2026-02-02T15:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:04 crc kubenswrapper[4733]: I0202 15:16:04.838494 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:04 crc kubenswrapper[4733]: I0202 15:16:04.838546 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:04 crc kubenswrapper[4733]: I0202 15:16:04.838566 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:04 crc kubenswrapper[4733]: I0202 15:16:04.838590 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:04 crc kubenswrapper[4733]: I0202 15:16:04.838607 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:04Z","lastTransitionTime":"2026-02-02T15:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:04 crc kubenswrapper[4733]: I0202 15:16:04.941850 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:04 crc kubenswrapper[4733]: I0202 15:16:04.941916 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:04 crc kubenswrapper[4733]: I0202 15:16:04.941933 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:04 crc kubenswrapper[4733]: I0202 15:16:04.941959 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:04 crc kubenswrapper[4733]: I0202 15:16:04.941976 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:04Z","lastTransitionTime":"2026-02-02T15:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.045099 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.045218 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.045246 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.045280 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.045304 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:05Z","lastTransitionTime":"2026-02-02T15:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.148744 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.148810 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.148826 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.148854 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.148877 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:05Z","lastTransitionTime":"2026-02-02T15:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.220886 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-12 21:34:04.394714433 +0000 UTC Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.252008 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.252074 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.252090 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.252108 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.252118 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:05Z","lastTransitionTime":"2026-02-02T15:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.254752 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.254758 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.254820 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:16:05 crc kubenswrapper[4733]: E0202 15:16:05.254987 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.255012 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:16:05 crc kubenswrapper[4733]: E0202 15:16:05.255116 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:16:05 crc kubenswrapper[4733]: E0202 15:16:05.255226 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:16:05 crc kubenswrapper[4733]: E0202 15:16:05.255316 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.354742 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.354777 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.354789 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.354804 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.354817 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:05Z","lastTransitionTime":"2026-02-02T15:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.457867 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.457942 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.457962 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.457986 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.458004 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:05Z","lastTransitionTime":"2026-02-02T15:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.561404 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.561476 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.561494 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.561520 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.561538 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:05Z","lastTransitionTime":"2026-02-02T15:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.664558 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.664619 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.664642 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.664673 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.664692 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:05Z","lastTransitionTime":"2026-02-02T15:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.767998 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.768047 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.768063 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.768085 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.768101 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:05Z","lastTransitionTime":"2026-02-02T15:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.873595 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.873671 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.873691 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.873719 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.873748 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:05Z","lastTransitionTime":"2026-02-02T15:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.977861 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.979886 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.979924 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.979953 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:05 crc kubenswrapper[4733]: I0202 15:16:05.979970 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:05Z","lastTransitionTime":"2026-02-02T15:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:06 crc kubenswrapper[4733]: I0202 15:16:06.083612 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:06 crc kubenswrapper[4733]: I0202 15:16:06.083677 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:06 crc kubenswrapper[4733]: I0202 15:16:06.083694 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:06 crc kubenswrapper[4733]: I0202 15:16:06.083719 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:06 crc kubenswrapper[4733]: I0202 15:16:06.083737 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:06Z","lastTransitionTime":"2026-02-02T15:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:06 crc kubenswrapper[4733]: I0202 15:16:06.187008 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:06 crc kubenswrapper[4733]: I0202 15:16:06.187068 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:06 crc kubenswrapper[4733]: I0202 15:16:06.187084 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:06 crc kubenswrapper[4733]: I0202 15:16:06.187108 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:06 crc kubenswrapper[4733]: I0202 15:16:06.187124 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:06Z","lastTransitionTime":"2026-02-02T15:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:06 crc kubenswrapper[4733]: I0202 15:16:06.222015 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 14:44:21.175959605 +0000 UTC Feb 02 15:16:06 crc kubenswrapper[4733]: I0202 15:16:06.290320 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:06 crc kubenswrapper[4733]: I0202 15:16:06.290693 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:06 crc kubenswrapper[4733]: I0202 15:16:06.290851 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:06 crc kubenswrapper[4733]: I0202 15:16:06.290995 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:06 crc kubenswrapper[4733]: I0202 15:16:06.291198 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:06Z","lastTransitionTime":"2026-02-02T15:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:06 crc kubenswrapper[4733]: I0202 15:16:06.394432 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:06 crc kubenswrapper[4733]: I0202 15:16:06.394503 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:06 crc kubenswrapper[4733]: I0202 15:16:06.394520 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:06 crc kubenswrapper[4733]: I0202 15:16:06.394549 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:06 crc kubenswrapper[4733]: I0202 15:16:06.394568 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:06Z","lastTransitionTime":"2026-02-02T15:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:06 crc kubenswrapper[4733]: I0202 15:16:06.497271 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:06 crc kubenswrapper[4733]: I0202 15:16:06.497612 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:06 crc kubenswrapper[4733]: I0202 15:16:06.497816 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:06 crc kubenswrapper[4733]: I0202 15:16:06.498032 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:06 crc kubenswrapper[4733]: I0202 15:16:06.498288 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:06Z","lastTransitionTime":"2026-02-02T15:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:06 crc kubenswrapper[4733]: I0202 15:16:06.601141 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:06 crc kubenswrapper[4733]: I0202 15:16:06.601299 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:06 crc kubenswrapper[4733]: I0202 15:16:06.601319 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:06 crc kubenswrapper[4733]: I0202 15:16:06.601346 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:06 crc kubenswrapper[4733]: I0202 15:16:06.601364 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:06Z","lastTransitionTime":"2026-02-02T15:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:06 crc kubenswrapper[4733]: I0202 15:16:06.704685 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:06 crc kubenswrapper[4733]: I0202 15:16:06.704754 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:06 crc kubenswrapper[4733]: I0202 15:16:06.704774 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:06 crc kubenswrapper[4733]: I0202 15:16:06.704841 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:06 crc kubenswrapper[4733]: I0202 15:16:06.704858 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:06Z","lastTransitionTime":"2026-02-02T15:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:06 crc kubenswrapper[4733]: I0202 15:16:06.807661 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:06 crc kubenswrapper[4733]: I0202 15:16:06.807785 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:06 crc kubenswrapper[4733]: I0202 15:16:06.807813 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:06 crc kubenswrapper[4733]: I0202 15:16:06.807840 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:06 crc kubenswrapper[4733]: I0202 15:16:06.807861 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:06Z","lastTransitionTime":"2026-02-02T15:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:06 crc kubenswrapper[4733]: I0202 15:16:06.911711 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:06 crc kubenswrapper[4733]: I0202 15:16:06.912219 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:06 crc kubenswrapper[4733]: I0202 15:16:06.912403 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:06 crc kubenswrapper[4733]: I0202 15:16:06.912609 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:06 crc kubenswrapper[4733]: I0202 15:16:06.912835 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:06Z","lastTransitionTime":"2026-02-02T15:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.016019 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.016080 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.016100 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.016124 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.016142 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:07Z","lastTransitionTime":"2026-02-02T15:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.112335 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.112388 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.112405 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.112429 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.112447 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:07Z","lastTransitionTime":"2026-02-02T15:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:07 crc kubenswrapper[4733]: E0202 15:16:07.132732 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd742269-802e-45ec-a88e-17aee17c574e\\\",\\\"systemUUID\\\":\\\"91b64b75-3cd9-4fb9-93f0-dbd3bcdcaec4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:07Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.137371 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.137471 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.137491 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.137515 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.137533 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:07Z","lastTransitionTime":"2026-02-02T15:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:07 crc kubenswrapper[4733]: E0202 15:16:07.157410 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd742269-802e-45ec-a88e-17aee17c574e\\\",\\\"systemUUID\\\":\\\"91b64b75-3cd9-4fb9-93f0-dbd3bcdcaec4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:07Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.162454 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.162692 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.162787 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.162884 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.162984 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:07Z","lastTransitionTime":"2026-02-02T15:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:07 crc kubenswrapper[4733]: E0202 15:16:07.181787 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd742269-802e-45ec-a88e-17aee17c574e\\\",\\\"systemUUID\\\":\\\"91b64b75-3cd9-4fb9-93f0-dbd3bcdcaec4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:07Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.187319 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.187393 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.187418 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.187449 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.187471 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:07Z","lastTransitionTime":"2026-02-02T15:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:07 crc kubenswrapper[4733]: E0202 15:16:07.202745 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd742269-802e-45ec-a88e-17aee17c574e\\\",\\\"systemUUID\\\":\\\"91b64b75-3cd9-4fb9-93f0-dbd3bcdcaec4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:07Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.207697 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.207759 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.207778 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.207806 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.207825 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:07Z","lastTransitionTime":"2026-02-02T15:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.222246 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 15:05:26.061454069 +0000 UTC Feb 02 15:16:07 crc kubenswrapper[4733]: E0202 15:16:07.231348 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd742269-802e-45ec-a88e-17aee17c574e\\\",\\\"systemUUID\\\":\\\"91b64b75-3cd9-4fb9-93f0-dbd3bcdcaec4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:07Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:07 crc kubenswrapper[4733]: E0202 15:16:07.231818 4733 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.233688 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.233751 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.233772 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.233798 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.233816 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:07Z","lastTransitionTime":"2026-02-02T15:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.254833 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.254841 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.254918 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.254971 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:16:07 crc kubenswrapper[4733]: E0202 15:16:07.255114 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:16:07 crc kubenswrapper[4733]: E0202 15:16:07.255249 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:16:07 crc kubenswrapper[4733]: E0202 15:16:07.257628 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:16:07 crc kubenswrapper[4733]: E0202 15:16:07.262689 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.285186 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:07Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.306268 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca57dabf1d1ee5c6fc29291b1be687ca7e97c72650084b02067057fccac41a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:07Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.326128 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:07Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.337505 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.337546 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.337558 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.337578 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.337593 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:07Z","lastTransitionTime":"2026-02-02T15:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.341842 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5xgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56681968-f264-440e-aa12-d4f81879efe9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e250891b1065a38a448aa6abd300b37543e819e87f006b03cc37983cb643982a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdjjl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5xgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:07Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.362378 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7r9fd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4dfd44b-b363-4162-9f43-9a8dc4a3b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7r9fd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:07Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.386808 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7082da2-f2e7-4b73-ba35-504e5cd4c920\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd383d69b9b27d52fa8672ff6e0294a74876d676973bdc0ca95125fef5f2d085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037e379a109ce230098e3ca223fd1dea3149de3ebe601f56e2bb295f3093a3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6fdd19dac506f2d5d7eedc44ec2cda92a9381b73e91a45d949e4fcde4dfc283\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13965f5abcfb0a36c33310518955f192673e630a1aaba94b49400026de5bc2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://992f4198192cbe4b1c89d575aa4a20c29c1115fd5eaa8986c6ed70c1d6b143c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0202 15:15:31.010033 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 15:15:31.011243 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3543528698/tls.crt::/tmp/serving-cert-3543528698/tls.key\\\\\\\"\\\\nI0202 15:15:36.416658 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 15:15:36.428087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 15:15:36.428134 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 15:15:36.428243 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 15:15:36.428280 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 15:15:36.439678 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 15:15:36.439710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439717 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439722 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 15:15:36.439726 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 15:15:36.439729 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 15:15:36.439732 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 15:15:36.439753 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 15:15:36.450716 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://226f5949820d4f498dbf6fa251c6edf1c6356900bd946e6afae2c29042b1c835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:07Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.409724 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da08f9aecc070b31df2ddc1be7c6eb2ad8ef68a931e57ca2e5c5302f4e75751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:07Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.426310 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8c597" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76cc68c3-9e90-45de-87b2-6ea1b68eadce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ac57831c1c880e64aaa7bedcf999133ee09fc5ecc5124b9d54ecedcc9a13c28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwr7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8c597\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:07Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.440013 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.440053 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.440064 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.440083 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.440097 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:07Z","lastTransitionTime":"2026-02-02T15:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.442679 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hbm26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45553dc5-54ea-4de9-914d-55b8fee477f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d112b1ea41c3d30a7cfa46d17db6aed0999eb0baaaf8a2f21c741d805a21451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d64j4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f76a1f655f871deb7e1364fcc5dbf6b564c4c386300bb49cb75f83518b64286e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d64j4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hbm26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:07Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.474097 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8206eca4-7d8b-4b59-a4c1-24631dd8cd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41d4a7bdc21f00b5039385d7227d8cb0f4ea3051d3b5065c71a49cb0bfad26a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c11b1859bf3bc0ec4d5e6a58d9becc1ac6333a1ed9820d83d634f1e93221c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c08b087d85be9bb83007c7e38a5de9182a5091716fb8a2ca5ded1b8191bebc40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fd4969f0f16c9aa44208d2bf578f77d9d50914a8cb1764507175aa2eb58e733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59175c8e5a7f3536168a69fd944bf3d536bcd7281eee96002343b7836a54401b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:07Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.495913 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nr6ps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e43cd902-791a-4042-bc08-cf03e075be32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2a89a5d1fc43696f33d313aee5dca316a249fdef6a23ff655e31f14cc449cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07e56782051c93b2a320aa539a42764b97647237d228bed2b6af48cee987c60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07e56782051c93b2a320aa539a42764b97647237d228bed2b6af48cee987c60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce850ac7bd4c2444b65d87232114d55b78d11c6f973537c615900092a45d55e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce850ac7bd4c2444b65d87232114d55b78d11c6f973537c615900092a45d55e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nr6ps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:07Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.521500 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5a81ada-e393-499a-aac6-7b40349958bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4acf2398e2e19bd54f482cb2273c54738e61627187860bb0b6daadddc2565b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831d85055203f153381eb51d98af0cd2e4c4a0309f837545ab3d855a680133e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1380122d1694c1286f1102e6396c6504b91f63c6eb157689dfc0ea7188c17f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5835cb53385359e9c58eaa5d4da2569c313e0fc265ad79a6c124e8cecb1a734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd6dc4bf5b9fd31bfbee6d9340fe981c0c5fca03a79cef9316340ff5a8b3d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db1db75e392ebea8b387b85cabc4276d553bf2c06694cf7f294a383909f32826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa356e9d1cf1236a6d325e762401b42848ed4c48bc4c538a5844e09de54d16c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa356e9d1cf1236a6d325e762401b42848ed4c48bc4c538a5844e09de54d16c1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T15:16:02Z\\\",\\\"message\\\":\\\"\\\\nI0202 15:16:02.301761 6382 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-5xgpn in node crc\\\\nI0202 15:16:02.301771 6382 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-8c597\\\\nI0202 15:16:02.301779 6382 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-5xgpn after 0 failed attempt(s)\\\\nI0202 15:16:02.301784 6382 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-8c597 in node crc\\\\nI0202 15:16:02.301793 6382 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-5xgpn\\\\nI0202 15:16:02.301794 6382 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-8c597 after 0 failed attempt(s)\\\\nI0202 15:16:02.301805 6382 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-8c597\\\\nI0202 15:16:02.301355 6382 services_controller.go:360] Finished syncing service metrics on namespace openshift-ingress-operator for network=default : 1.801637ms\\\\nI0202 15:16:02.301684 6382 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI0202 15:16:02.301833 6382 services_controller.go:356] Processing sync for service openshift-console-operator/metrics for network=default\\\\nI0202 15:16:02\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:16:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qxvkf_openshift-ovn-kubernetes(f5a81ada-e393-499a-aac6-7b40349958bb)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efaf2216e4c1750139f080a93b4b608737dab2883de5fddfd9a99248d517b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxvkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:07Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.543442 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.543517 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.543542 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.543574 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.543599 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:07Z","lastTransitionTime":"2026-02-02T15:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.543628 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"402a6dc5-6bc7-4d7c-a909-9bb4ca48365d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d5b032497b592363a03f89cb31ba1b551f479004072dddffce540c41190ab81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a3fd02d55a10647428a05a9f53a1429a9827059c523d5deab3c4581c8f1b933\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11600ada13fbc7db1831de385202bc1f041efb656cc58aa49b0f71fe902db9bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fd39f5a05dec7a756dfd519f5df3ef7c77bbb5b8633c5af18a09ff8029cb079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:07Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.567551 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83d0e810-ea5c-47aa-947b-7865b045bf94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6144eb32f88bc106f7724e48133b568a137c4f3fbbfe30c7fcef405d3ab735a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51318ad71b42f4cc8d376aa50e44580e68da89d83d80defb996329a2ee5c2ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zgsv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:07Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.589320 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:07Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.608744 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b38b8410bea95465ee58803c4160039bf7c082ca47893680da7469418f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27003b336112760daa2079f4819641850c27fbc8415b63b5094c86195291aca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:07Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.631263 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jh6vk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76a05d6-d9c7-4183-ad90-b3131f225841\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f2659e2a914d3191daddbfb5ad109e487f03dabb559d324244e127add7d41d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jh6vk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:07Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.646769 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.647046 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.647242 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.647484 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.647627 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:07Z","lastTransitionTime":"2026-02-02T15:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.751270 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.751328 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.751346 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.751373 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.751392 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:07Z","lastTransitionTime":"2026-02-02T15:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.857712 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.858472 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.858501 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.858529 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.858547 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:07Z","lastTransitionTime":"2026-02-02T15:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.962115 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.962211 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.962230 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.962255 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:07 crc kubenswrapper[4733]: I0202 15:16:07.962275 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:07Z","lastTransitionTime":"2026-02-02T15:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:08 crc kubenswrapper[4733]: I0202 15:16:08.065557 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:08 crc kubenswrapper[4733]: I0202 15:16:08.066364 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:08 crc kubenswrapper[4733]: I0202 15:16:08.066415 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:08 crc kubenswrapper[4733]: I0202 15:16:08.066451 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:08 crc kubenswrapper[4733]: I0202 15:16:08.066471 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:08Z","lastTransitionTime":"2026-02-02T15:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:08 crc kubenswrapper[4733]: I0202 15:16:08.169844 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:08 crc kubenswrapper[4733]: I0202 15:16:08.169909 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:08 crc kubenswrapper[4733]: I0202 15:16:08.169933 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:08 crc kubenswrapper[4733]: I0202 15:16:08.169964 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:08 crc kubenswrapper[4733]: I0202 15:16:08.169987 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:08Z","lastTransitionTime":"2026-02-02T15:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:08 crc kubenswrapper[4733]: I0202 15:16:08.222537 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-11 18:27:38.932908224 +0000 UTC Feb 02 15:16:08 crc kubenswrapper[4733]: I0202 15:16:08.273144 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:08 crc kubenswrapper[4733]: I0202 15:16:08.273267 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:08 crc kubenswrapper[4733]: I0202 15:16:08.273294 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:08 crc kubenswrapper[4733]: I0202 15:16:08.273326 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:08 crc kubenswrapper[4733]: I0202 15:16:08.273349 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:08Z","lastTransitionTime":"2026-02-02T15:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:08 crc kubenswrapper[4733]: I0202 15:16:08.376892 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:08 crc kubenswrapper[4733]: I0202 15:16:08.376952 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:08 crc kubenswrapper[4733]: I0202 15:16:08.376969 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:08 crc kubenswrapper[4733]: I0202 15:16:08.376991 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:08 crc kubenswrapper[4733]: I0202 15:16:08.377008 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:08Z","lastTransitionTime":"2026-02-02T15:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:08 crc kubenswrapper[4733]: I0202 15:16:08.481359 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:08 crc kubenswrapper[4733]: I0202 15:16:08.481984 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:08 crc kubenswrapper[4733]: I0202 15:16:08.482006 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:08 crc kubenswrapper[4733]: I0202 15:16:08.482030 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:08 crc kubenswrapper[4733]: I0202 15:16:08.482047 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:08Z","lastTransitionTime":"2026-02-02T15:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:08 crc kubenswrapper[4733]: I0202 15:16:08.585723 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:08 crc kubenswrapper[4733]: I0202 15:16:08.585834 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:08 crc kubenswrapper[4733]: I0202 15:16:08.585859 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:08 crc kubenswrapper[4733]: I0202 15:16:08.585888 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:08 crc kubenswrapper[4733]: I0202 15:16:08.585910 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:08Z","lastTransitionTime":"2026-02-02T15:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:08 crc kubenswrapper[4733]: I0202 15:16:08.689037 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:08 crc kubenswrapper[4733]: I0202 15:16:08.689108 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:08 crc kubenswrapper[4733]: I0202 15:16:08.689122 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:08 crc kubenswrapper[4733]: I0202 15:16:08.689196 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:08 crc kubenswrapper[4733]: I0202 15:16:08.689217 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:08Z","lastTransitionTime":"2026-02-02T15:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:08 crc kubenswrapper[4733]: I0202 15:16:08.792885 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:08 crc kubenswrapper[4733]: I0202 15:16:08.792963 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:08 crc kubenswrapper[4733]: I0202 15:16:08.792987 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:08 crc kubenswrapper[4733]: I0202 15:16:08.793019 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:08 crc kubenswrapper[4733]: I0202 15:16:08.793045 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:08Z","lastTransitionTime":"2026-02-02T15:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:08 crc kubenswrapper[4733]: I0202 15:16:08.897146 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:08 crc kubenswrapper[4733]: I0202 15:16:08.897236 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:08 crc kubenswrapper[4733]: I0202 15:16:08.897255 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:08 crc kubenswrapper[4733]: I0202 15:16:08.897275 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:08 crc kubenswrapper[4733]: I0202 15:16:08.897290 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:08Z","lastTransitionTime":"2026-02-02T15:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:08 crc kubenswrapper[4733]: I0202 15:16:08.934754 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 15:16:08 crc kubenswrapper[4733]: I0202 15:16:08.934938 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:16:08 crc kubenswrapper[4733]: E0202 15:16:08.935054 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 15:16:40.935017325 +0000 UTC m=+84.386478723 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:16:08 crc kubenswrapper[4733]: E0202 15:16:08.935064 4733 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 02 15:16:08 crc kubenswrapper[4733]: E0202 15:16:08.935226 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-02 15:16:40.93519764 +0000 UTC m=+84.386659038 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.000200 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.000249 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.000264 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.000284 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.000299 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:09Z","lastTransitionTime":"2026-02-02T15:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.036085 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d4dfd44b-b363-4162-9f43-9a8dc4a3b110-metrics-certs\") pod \"network-metrics-daemon-7r9fd\" (UID: \"d4dfd44b-b363-4162-9f43-9a8dc4a3b110\") " pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.036156 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.036231 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.036277 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:16:09 crc kubenswrapper[4733]: E0202 15:16:09.036344 4733 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 02 15:16:09 crc kubenswrapper[4733]: E0202 15:16:09.036392 4733 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 02 15:16:09 crc kubenswrapper[4733]: E0202 15:16:09.036467 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 02 15:16:09 crc kubenswrapper[4733]: E0202 15:16:09.036506 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 02 15:16:09 crc kubenswrapper[4733]: E0202 15:16:09.036527 4733 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 15:16:09 crc kubenswrapper[4733]: E0202 15:16:09.036470 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d4dfd44b-b363-4162-9f43-9a8dc4a3b110-metrics-certs podName:d4dfd44b-b363-4162-9f43-9a8dc4a3b110 nodeName:}" failed. No retries permitted until 2026-02-02 15:16:41.036433466 +0000 UTC m=+84.487894914 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d4dfd44b-b363-4162-9f43-9a8dc4a3b110-metrics-certs") pod "network-metrics-daemon-7r9fd" (UID: "d4dfd44b-b363-4162-9f43-9a8dc4a3b110") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 02 15:16:09 crc kubenswrapper[4733]: E0202 15:16:09.036608 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-02 15:16:41.03658319 +0000 UTC m=+84.488044578 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 02 15:16:09 crc kubenswrapper[4733]: E0202 15:16:09.036469 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 02 15:16:09 crc kubenswrapper[4733]: E0202 15:16:09.036671 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 02 15:16:09 crc kubenswrapper[4733]: E0202 15:16:09.036694 4733 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 15:16:09 crc kubenswrapper[4733]: E0202 15:16:09.036630 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-02 15:16:41.036618131 +0000 UTC m=+84.488079519 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 15:16:09 crc kubenswrapper[4733]: E0202 15:16:09.036781 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-02 15:16:41.036753365 +0000 UTC m=+84.488214763 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.103779 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.103842 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.103859 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.103888 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.103907 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:09Z","lastTransitionTime":"2026-02-02T15:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.207181 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.207241 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.207257 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.207281 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.207298 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:09Z","lastTransitionTime":"2026-02-02T15:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.222891 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 12:11:34.451217159 +0000 UTC Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.255362 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.255452 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.255485 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.255369 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:16:09 crc kubenswrapper[4733]: E0202 15:16:09.255862 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:16:09 crc kubenswrapper[4733]: E0202 15:16:09.255985 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:16:09 crc kubenswrapper[4733]: E0202 15:16:09.256067 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:16:09 crc kubenswrapper[4733]: E0202 15:16:09.256322 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.309990 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.310534 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.310573 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.310596 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.310614 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:09Z","lastTransitionTime":"2026-02-02T15:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.413546 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.413617 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.413639 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.413667 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.413684 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:09Z","lastTransitionTime":"2026-02-02T15:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.516783 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.516855 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.516873 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.516898 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.516915 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:09Z","lastTransitionTime":"2026-02-02T15:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.619603 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.619709 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.619728 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.619752 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.619768 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:09Z","lastTransitionTime":"2026-02-02T15:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.723308 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.723366 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.723390 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.723415 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.723431 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:09Z","lastTransitionTime":"2026-02-02T15:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.826203 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.826263 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.826280 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.826304 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.826322 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:09Z","lastTransitionTime":"2026-02-02T15:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.929597 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.929663 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.929681 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.929707 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:09 crc kubenswrapper[4733]: I0202 15:16:09.929727 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:09Z","lastTransitionTime":"2026-02-02T15:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:10 crc kubenswrapper[4733]: I0202 15:16:10.032997 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:10 crc kubenswrapper[4733]: I0202 15:16:10.033057 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:10 crc kubenswrapper[4733]: I0202 15:16:10.033075 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:10 crc kubenswrapper[4733]: I0202 15:16:10.033099 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:10 crc kubenswrapper[4733]: I0202 15:16:10.033117 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:10Z","lastTransitionTime":"2026-02-02T15:16:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:10 crc kubenswrapper[4733]: I0202 15:16:10.136274 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:10 crc kubenswrapper[4733]: I0202 15:16:10.136332 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:10 crc kubenswrapper[4733]: I0202 15:16:10.136350 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:10 crc kubenswrapper[4733]: I0202 15:16:10.136374 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:10 crc kubenswrapper[4733]: I0202 15:16:10.136390 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:10Z","lastTransitionTime":"2026-02-02T15:16:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:10 crc kubenswrapper[4733]: I0202 15:16:10.223509 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-27 17:11:15.072495781 +0000 UTC Feb 02 15:16:10 crc kubenswrapper[4733]: I0202 15:16:10.239760 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:10 crc kubenswrapper[4733]: I0202 15:16:10.239814 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:10 crc kubenswrapper[4733]: I0202 15:16:10.239835 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:10 crc kubenswrapper[4733]: I0202 15:16:10.239863 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:10 crc kubenswrapper[4733]: I0202 15:16:10.239884 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:10Z","lastTransitionTime":"2026-02-02T15:16:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:10 crc kubenswrapper[4733]: I0202 15:16:10.343816 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:10 crc kubenswrapper[4733]: I0202 15:16:10.343886 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:10 crc kubenswrapper[4733]: I0202 15:16:10.343907 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:10 crc kubenswrapper[4733]: I0202 15:16:10.343931 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:10 crc kubenswrapper[4733]: I0202 15:16:10.343949 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:10Z","lastTransitionTime":"2026-02-02T15:16:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:10 crc kubenswrapper[4733]: I0202 15:16:10.446250 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:10 crc kubenswrapper[4733]: I0202 15:16:10.446278 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:10 crc kubenswrapper[4733]: I0202 15:16:10.446286 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:10 crc kubenswrapper[4733]: I0202 15:16:10.446300 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:10 crc kubenswrapper[4733]: I0202 15:16:10.446307 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:10Z","lastTransitionTime":"2026-02-02T15:16:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:10 crc kubenswrapper[4733]: I0202 15:16:10.548491 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:10 crc kubenswrapper[4733]: I0202 15:16:10.548545 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:10 crc kubenswrapper[4733]: I0202 15:16:10.548561 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:10 crc kubenswrapper[4733]: I0202 15:16:10.548583 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:10 crc kubenswrapper[4733]: I0202 15:16:10.548609 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:10Z","lastTransitionTime":"2026-02-02T15:16:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:10 crc kubenswrapper[4733]: I0202 15:16:10.651149 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:10 crc kubenswrapper[4733]: I0202 15:16:10.651249 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:10 crc kubenswrapper[4733]: I0202 15:16:10.651268 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:10 crc kubenswrapper[4733]: I0202 15:16:10.651292 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:10 crc kubenswrapper[4733]: I0202 15:16:10.651360 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:10Z","lastTransitionTime":"2026-02-02T15:16:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:10 crc kubenswrapper[4733]: I0202 15:16:10.755042 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:10 crc kubenswrapper[4733]: I0202 15:16:10.755090 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:10 crc kubenswrapper[4733]: I0202 15:16:10.755104 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:10 crc kubenswrapper[4733]: I0202 15:16:10.755124 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:10 crc kubenswrapper[4733]: I0202 15:16:10.755138 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:10Z","lastTransitionTime":"2026-02-02T15:16:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:10 crc kubenswrapper[4733]: I0202 15:16:10.858342 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:10 crc kubenswrapper[4733]: I0202 15:16:10.858417 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:10 crc kubenswrapper[4733]: I0202 15:16:10.858442 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:10 crc kubenswrapper[4733]: I0202 15:16:10.858474 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:10 crc kubenswrapper[4733]: I0202 15:16:10.858497 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:10Z","lastTransitionTime":"2026-02-02T15:16:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:10 crc kubenswrapper[4733]: I0202 15:16:10.961541 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:10 crc kubenswrapper[4733]: I0202 15:16:10.961601 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:10 crc kubenswrapper[4733]: I0202 15:16:10.961620 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:10 crc kubenswrapper[4733]: I0202 15:16:10.961644 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:10 crc kubenswrapper[4733]: I0202 15:16:10.961661 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:10Z","lastTransitionTime":"2026-02-02T15:16:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.064938 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.065013 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.065037 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.065069 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.065093 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:11Z","lastTransitionTime":"2026-02-02T15:16:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.167670 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.167727 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.167744 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.167768 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.167785 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:11Z","lastTransitionTime":"2026-02-02T15:16:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.223913 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 02:57:55.412596653 +0000 UTC Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.254392 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.254484 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.254485 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:16:11 crc kubenswrapper[4733]: E0202 15:16:11.254605 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.254741 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:16:11 crc kubenswrapper[4733]: E0202 15:16:11.254799 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:16:11 crc kubenswrapper[4733]: E0202 15:16:11.254904 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:16:11 crc kubenswrapper[4733]: E0202 15:16:11.255074 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.270358 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.270407 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.270425 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.270448 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.270466 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:11Z","lastTransitionTime":"2026-02-02T15:16:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.373533 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.373591 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.373611 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.373636 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.373653 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:11Z","lastTransitionTime":"2026-02-02T15:16:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.477558 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.477629 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.477653 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.477684 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.477706 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:11Z","lastTransitionTime":"2026-02-02T15:16:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.580581 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.580655 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.580678 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.580711 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.580739 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:11Z","lastTransitionTime":"2026-02-02T15:16:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.683691 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.683750 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.683773 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.683804 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.683827 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:11Z","lastTransitionTime":"2026-02-02T15:16:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.786864 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.786925 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.786942 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.786965 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.786982 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:11Z","lastTransitionTime":"2026-02-02T15:16:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.890291 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.890357 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.890379 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.890409 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.890431 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:11Z","lastTransitionTime":"2026-02-02T15:16:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.993309 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.993383 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.993406 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.993434 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:11 crc kubenswrapper[4733]: I0202 15:16:11.993457 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:11Z","lastTransitionTime":"2026-02-02T15:16:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:12 crc kubenswrapper[4733]: I0202 15:16:12.096461 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:12 crc kubenswrapper[4733]: I0202 15:16:12.096523 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:12 crc kubenswrapper[4733]: I0202 15:16:12.096541 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:12 crc kubenswrapper[4733]: I0202 15:16:12.096563 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:12 crc kubenswrapper[4733]: I0202 15:16:12.096580 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:12Z","lastTransitionTime":"2026-02-02T15:16:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:12 crc kubenswrapper[4733]: I0202 15:16:12.199211 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:12 crc kubenswrapper[4733]: I0202 15:16:12.199257 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:12 crc kubenswrapper[4733]: I0202 15:16:12.199271 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:12 crc kubenswrapper[4733]: I0202 15:16:12.199292 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:12 crc kubenswrapper[4733]: I0202 15:16:12.199309 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:12Z","lastTransitionTime":"2026-02-02T15:16:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:12 crc kubenswrapper[4733]: I0202 15:16:12.224901 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-20 07:43:14.451300555 +0000 UTC Feb 02 15:16:12 crc kubenswrapper[4733]: I0202 15:16:12.307619 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:12 crc kubenswrapper[4733]: I0202 15:16:12.307677 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:12 crc kubenswrapper[4733]: I0202 15:16:12.307695 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:12 crc kubenswrapper[4733]: I0202 15:16:12.307718 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:12 crc kubenswrapper[4733]: I0202 15:16:12.307759 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:12Z","lastTransitionTime":"2026-02-02T15:16:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:12 crc kubenswrapper[4733]: I0202 15:16:12.410273 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:12 crc kubenswrapper[4733]: I0202 15:16:12.410322 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:12 crc kubenswrapper[4733]: I0202 15:16:12.410338 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:12 crc kubenswrapper[4733]: I0202 15:16:12.410359 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:12 crc kubenswrapper[4733]: I0202 15:16:12.410376 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:12Z","lastTransitionTime":"2026-02-02T15:16:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:12 crc kubenswrapper[4733]: I0202 15:16:12.512442 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:12 crc kubenswrapper[4733]: I0202 15:16:12.512500 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:12 crc kubenswrapper[4733]: I0202 15:16:12.512517 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:12 crc kubenswrapper[4733]: I0202 15:16:12.512539 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:12 crc kubenswrapper[4733]: I0202 15:16:12.512558 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:12Z","lastTransitionTime":"2026-02-02T15:16:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:12 crc kubenswrapper[4733]: I0202 15:16:12.616205 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:12 crc kubenswrapper[4733]: I0202 15:16:12.616262 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:12 crc kubenswrapper[4733]: I0202 15:16:12.616279 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:12 crc kubenswrapper[4733]: I0202 15:16:12.616305 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:12 crc kubenswrapper[4733]: I0202 15:16:12.616321 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:12Z","lastTransitionTime":"2026-02-02T15:16:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:12 crc kubenswrapper[4733]: I0202 15:16:12.719780 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:12 crc kubenswrapper[4733]: I0202 15:16:12.719840 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:12 crc kubenswrapper[4733]: I0202 15:16:12.719857 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:12 crc kubenswrapper[4733]: I0202 15:16:12.720356 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:12 crc kubenswrapper[4733]: I0202 15:16:12.720412 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:12Z","lastTransitionTime":"2026-02-02T15:16:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:12 crc kubenswrapper[4733]: I0202 15:16:12.824046 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:12 crc kubenswrapper[4733]: I0202 15:16:12.824108 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:12 crc kubenswrapper[4733]: I0202 15:16:12.824127 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:12 crc kubenswrapper[4733]: I0202 15:16:12.824151 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:12 crc kubenswrapper[4733]: I0202 15:16:12.824196 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:12Z","lastTransitionTime":"2026-02-02T15:16:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:12 crc kubenswrapper[4733]: I0202 15:16:12.927542 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:12 crc kubenswrapper[4733]: I0202 15:16:12.927589 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:12 crc kubenswrapper[4733]: I0202 15:16:12.927606 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:12 crc kubenswrapper[4733]: I0202 15:16:12.927627 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:12 crc kubenswrapper[4733]: I0202 15:16:12.927644 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:12Z","lastTransitionTime":"2026-02-02T15:16:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.030777 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.030833 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.030850 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.030872 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.030889 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:13Z","lastTransitionTime":"2026-02-02T15:16:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.134218 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.134286 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.134308 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.134337 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.134358 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:13Z","lastTransitionTime":"2026-02-02T15:16:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.225729 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 01:23:11.966096297 +0000 UTC Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.237043 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.237120 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.237142 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.237196 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.237215 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:13Z","lastTransitionTime":"2026-02-02T15:16:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.254444 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.254468 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.254518 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.254555 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:16:13 crc kubenswrapper[4733]: E0202 15:16:13.254758 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:16:13 crc kubenswrapper[4733]: E0202 15:16:13.254913 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:16:13 crc kubenswrapper[4733]: E0202 15:16:13.255072 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:16:13 crc kubenswrapper[4733]: E0202 15:16:13.255200 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.339632 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.339967 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.340105 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.340330 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.340508 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:13Z","lastTransitionTime":"2026-02-02T15:16:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.444305 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.444399 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.444417 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.444441 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.444459 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:13Z","lastTransitionTime":"2026-02-02T15:16:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.547198 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.547261 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.547283 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.547312 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.547332 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:13Z","lastTransitionTime":"2026-02-02T15:16:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.649988 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.650042 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.650059 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.650082 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.650101 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:13Z","lastTransitionTime":"2026-02-02T15:16:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.753068 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.753130 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.753152 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.753205 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.753227 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:13Z","lastTransitionTime":"2026-02-02T15:16:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.856775 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.856840 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.856858 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.856883 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.856900 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:13Z","lastTransitionTime":"2026-02-02T15:16:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.954310 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.961001 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.961090 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.961113 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.961148 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.961230 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:13Z","lastTransitionTime":"2026-02-02T15:16:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.974748 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.979015 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca57dabf1d1ee5c6fc29291b1be687ca7e97c72650084b02067057fccac41a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:13Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:13 crc kubenswrapper[4733]: I0202 15:16:13.999756 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:13Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.017718 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5xgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56681968-f264-440e-aa12-d4f81879efe9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e250891b1065a38a448aa6abd300b37543e819e87f006b03cc37983cb643982a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdjjl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5xgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:14Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.035356 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7r9fd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4dfd44b-b363-4162-9f43-9a8dc4a3b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7r9fd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:14Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.056838 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7082da2-f2e7-4b73-ba35-504e5cd4c920\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd383d69b9b27d52fa8672ff6e0294a74876d676973bdc0ca95125fef5f2d085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037e379a109ce230098e3ca223fd1dea3149de3ebe601f56e2bb295f3093a3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6fdd19dac506f2d5d7eedc44ec2cda92a9381b73e91a45d949e4fcde4dfc283\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13965f5abcfb0a36c33310518955f192673e630a1aaba94b49400026de5bc2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://992f4198192cbe4b1c89d575aa4a20c29c1115fd5eaa8986c6ed70c1d6b143c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0202 15:15:31.010033 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 15:15:31.011243 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3543528698/tls.crt::/tmp/serving-cert-3543528698/tls.key\\\\\\\"\\\\nI0202 15:15:36.416658 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 15:15:36.428087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 15:15:36.428134 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 15:15:36.428243 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 15:15:36.428280 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 15:15:36.439678 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 15:15:36.439710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439717 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439722 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 15:15:36.439726 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 15:15:36.439729 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 15:15:36.439732 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 15:15:36.439753 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 15:15:36.450716 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://226f5949820d4f498dbf6fa251c6edf1c6356900bd946e6afae2c29042b1c835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:14Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.063990 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.064045 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.064069 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.064098 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.064122 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:14Z","lastTransitionTime":"2026-02-02T15:16:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.077927 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da08f9aecc070b31df2ddc1be7c6eb2ad8ef68a931e57ca2e5c5302f4e75751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:14Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.093250 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:14Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.108659 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hbm26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45553dc5-54ea-4de9-914d-55b8fee477f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d112b1ea41c3d30a7cfa46d17db6aed0999eb0baaaf8a2f21c741d805a21451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d64j4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f76a1f655f871deb7e1364fcc5dbf6b564c4c386300bb49cb75f83518b64286e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d64j4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hbm26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:14Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.140685 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8206eca4-7d8b-4b59-a4c1-24631dd8cd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41d4a7bdc21f00b5039385d7227d8cb0f4ea3051d3b5065c71a49cb0bfad26a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c11b1859bf3bc0ec4d5e6a58d9becc1ac6333a1ed9820d83d634f1e93221c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c08b087d85be9bb83007c7e38a5de9182a5091716fb8a2ca5ded1b8191bebc40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fd4969f0f16c9aa44208d2bf578f77d9d50914a8cb1764507175aa2eb58e733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59175c8e5a7f3536168a69fd944bf3d536bcd7281eee96002343b7836a54401b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:14Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.166819 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.166886 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.166905 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.166932 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.166950 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:14Z","lastTransitionTime":"2026-02-02T15:16:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.168983 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nr6ps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e43cd902-791a-4042-bc08-cf03e075be32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2a89a5d1fc43696f33d313aee5dca316a249fdef6a23ff655e31f14cc449cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07e56782051c93b2a320aa539a42764b97647237d228bed2b6af48cee987c60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07e56782051c93b2a320aa539a42764b97647237d228bed2b6af48cee987c60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce850ac7bd4c2444b65d87232114d55b78d11c6f973537c615900092a45d55e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce850ac7bd4c2444b65d87232114d55b78d11c6f973537c615900092a45d55e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nr6ps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:14Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.200955 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5a81ada-e393-499a-aac6-7b40349958bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4acf2398e2e19bd54f482cb2273c54738e61627187860bb0b6daadddc2565b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831d85055203f153381eb51d98af0cd2e4c4a0309f837545ab3d855a680133e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1380122d1694c1286f1102e6396c6504b91f63c6eb157689dfc0ea7188c17f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5835cb53385359e9c58eaa5d4da2569c313e0fc265ad79a6c124e8cecb1a734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd6dc4bf5b9fd31bfbee6d9340fe981c0c5fca03a79cef9316340ff5a8b3d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db1db75e392ebea8b387b85cabc4276d553bf2c06694cf7f294a383909f32826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa356e9d1cf1236a6d325e762401b42848ed4c48bc4c538a5844e09de54d16c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa356e9d1cf1236a6d325e762401b42848ed4c48bc4c538a5844e09de54d16c1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T15:16:02Z\\\",\\\"message\\\":\\\"\\\\nI0202 15:16:02.301761 6382 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-5xgpn in node crc\\\\nI0202 15:16:02.301771 6382 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-8c597\\\\nI0202 15:16:02.301779 6382 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-5xgpn after 0 failed attempt(s)\\\\nI0202 15:16:02.301784 6382 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-8c597 in node crc\\\\nI0202 15:16:02.301793 6382 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-5xgpn\\\\nI0202 15:16:02.301794 6382 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-8c597 after 0 failed attempt(s)\\\\nI0202 15:16:02.301805 6382 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-8c597\\\\nI0202 15:16:02.301355 6382 services_controller.go:360] Finished syncing service metrics on namespace openshift-ingress-operator for network=default : 1.801637ms\\\\nI0202 15:16:02.301684 6382 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI0202 15:16:02.301833 6382 services_controller.go:356] Processing sync for service openshift-console-operator/metrics for network=default\\\\nI0202 15:16:02\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:16:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qxvkf_openshift-ovn-kubernetes(f5a81ada-e393-499a-aac6-7b40349958bb)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efaf2216e4c1750139f080a93b4b608737dab2883de5fddfd9a99248d517b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxvkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:14Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.215955 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8c597" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76cc68c3-9e90-45de-87b2-6ea1b68eadce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ac57831c1c880e64aaa7bedcf999133ee09fc5ecc5124b9d54ecedcc9a13c28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwr7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8c597\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:14Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.226796 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-16 23:02:58.517455637 +0000 UTC Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.235600 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"402a6dc5-6bc7-4d7c-a909-9bb4ca48365d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d5b032497b592363a03f89cb31ba1b551f479004072dddffce540c41190ab81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a3fd02d55a10647428a05a9f53a1429a9827059c523d5deab3c4581c8f1b933\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11600ada13fbc7db1831de385202bc1f041efb656cc58aa49b0f71fe902db9bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fd39f5a05dec7a756dfd519f5df3ef7c77bbb5b8633c5af18a09ff8029cb079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:14Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.255579 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:14Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.272990 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.273050 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.273069 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.273096 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.273129 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:14Z","lastTransitionTime":"2026-02-02T15:16:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.275602 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b38b8410bea95465ee58803c4160039bf7c082ca47893680da7469418f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27003b336112760daa2079f4819641850c27fbc8415b63b5094c86195291aca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:14Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.295843 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jh6vk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76a05d6-d9c7-4183-ad90-b3131f225841\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f2659e2a914d3191daddbfb5ad109e487f03dabb559d324244e127add7d41d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jh6vk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:14Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.313742 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83d0e810-ea5c-47aa-947b-7865b045bf94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6144eb32f88bc106f7724e48133b568a137c4f3fbbfe30c7fcef405d3ab735a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51318ad71b42f4cc8d376aa50e44580e68da89d83d80defb996329a2ee5c2ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zgsv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:14Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.376577 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.376638 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.376656 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.376680 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.376697 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:14Z","lastTransitionTime":"2026-02-02T15:16:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.479876 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.479948 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.479965 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.479988 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.480009 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:14Z","lastTransitionTime":"2026-02-02T15:16:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.583462 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.583516 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.583538 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.583566 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.583587 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:14Z","lastTransitionTime":"2026-02-02T15:16:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.685944 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.685977 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.685987 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.686003 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.686015 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:14Z","lastTransitionTime":"2026-02-02T15:16:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.789199 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.789256 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.789274 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.789298 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.789316 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:14Z","lastTransitionTime":"2026-02-02T15:16:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.892418 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.892490 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.892509 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.892537 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.892554 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:14Z","lastTransitionTime":"2026-02-02T15:16:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.995463 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.995507 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.995519 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.995538 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:14 crc kubenswrapper[4733]: I0202 15:16:14.995550 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:14Z","lastTransitionTime":"2026-02-02T15:16:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:15 crc kubenswrapper[4733]: I0202 15:16:15.098503 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:15 crc kubenswrapper[4733]: I0202 15:16:15.098562 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:15 crc kubenswrapper[4733]: I0202 15:16:15.098578 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:15 crc kubenswrapper[4733]: I0202 15:16:15.098598 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:15 crc kubenswrapper[4733]: I0202 15:16:15.098614 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:15Z","lastTransitionTime":"2026-02-02T15:16:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:15 crc kubenswrapper[4733]: I0202 15:16:15.202004 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:15 crc kubenswrapper[4733]: I0202 15:16:15.202071 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:15 crc kubenswrapper[4733]: I0202 15:16:15.202100 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:15 crc kubenswrapper[4733]: I0202 15:16:15.202129 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:15 crc kubenswrapper[4733]: I0202 15:16:15.202148 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:15Z","lastTransitionTime":"2026-02-02T15:16:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:15 crc kubenswrapper[4733]: I0202 15:16:15.226925 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-02 06:42:49.269490812 +0000 UTC Feb 02 15:16:15 crc kubenswrapper[4733]: I0202 15:16:15.254424 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:16:15 crc kubenswrapper[4733]: I0202 15:16:15.254405 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:16:15 crc kubenswrapper[4733]: I0202 15:16:15.254494 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:16:15 crc kubenswrapper[4733]: I0202 15:16:15.254537 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:16:15 crc kubenswrapper[4733]: E0202 15:16:15.254680 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:16:15 crc kubenswrapper[4733]: E0202 15:16:15.254806 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:16:15 crc kubenswrapper[4733]: E0202 15:16:15.254916 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:16:15 crc kubenswrapper[4733]: E0202 15:16:15.254995 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:16:15 crc kubenswrapper[4733]: I0202 15:16:15.256472 4733 scope.go:117] "RemoveContainer" containerID="fa356e9d1cf1236a6d325e762401b42848ed4c48bc4c538a5844e09de54d16c1" Feb 02 15:16:15 crc kubenswrapper[4733]: E0202 15:16:15.256954 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qxvkf_openshift-ovn-kubernetes(f5a81ada-e393-499a-aac6-7b40349958bb)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" Feb 02 15:16:15 crc kubenswrapper[4733]: I0202 15:16:15.304632 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:15 crc kubenswrapper[4733]: I0202 15:16:15.304680 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:15 crc kubenswrapper[4733]: I0202 15:16:15.304697 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:15 crc kubenswrapper[4733]: I0202 15:16:15.304721 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:15 crc kubenswrapper[4733]: I0202 15:16:15.304739 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:15Z","lastTransitionTime":"2026-02-02T15:16:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:15 crc kubenswrapper[4733]: I0202 15:16:15.407528 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:15 crc kubenswrapper[4733]: I0202 15:16:15.407586 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:15 crc kubenswrapper[4733]: I0202 15:16:15.407603 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:15 crc kubenswrapper[4733]: I0202 15:16:15.407627 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:15 crc kubenswrapper[4733]: I0202 15:16:15.407643 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:15Z","lastTransitionTime":"2026-02-02T15:16:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:15 crc kubenswrapper[4733]: I0202 15:16:15.510708 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:15 crc kubenswrapper[4733]: I0202 15:16:15.510762 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:15 crc kubenswrapper[4733]: I0202 15:16:15.510778 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:15 crc kubenswrapper[4733]: I0202 15:16:15.510803 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:15 crc kubenswrapper[4733]: I0202 15:16:15.510825 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:15Z","lastTransitionTime":"2026-02-02T15:16:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:15 crc kubenswrapper[4733]: I0202 15:16:15.613617 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:15 crc kubenswrapper[4733]: I0202 15:16:15.613721 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:15 crc kubenswrapper[4733]: I0202 15:16:15.613738 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:15 crc kubenswrapper[4733]: I0202 15:16:15.613765 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:15 crc kubenswrapper[4733]: I0202 15:16:15.613787 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:15Z","lastTransitionTime":"2026-02-02T15:16:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:15 crc kubenswrapper[4733]: I0202 15:16:15.716918 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:15 crc kubenswrapper[4733]: I0202 15:16:15.716977 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:15 crc kubenswrapper[4733]: I0202 15:16:15.716996 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:15 crc kubenswrapper[4733]: I0202 15:16:15.717020 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:15 crc kubenswrapper[4733]: I0202 15:16:15.717043 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:15Z","lastTransitionTime":"2026-02-02T15:16:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:15 crc kubenswrapper[4733]: I0202 15:16:15.819850 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:15 crc kubenswrapper[4733]: I0202 15:16:15.819958 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:15 crc kubenswrapper[4733]: I0202 15:16:15.819980 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:15 crc kubenswrapper[4733]: I0202 15:16:15.820009 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:15 crc kubenswrapper[4733]: I0202 15:16:15.820030 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:15Z","lastTransitionTime":"2026-02-02T15:16:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:15 crc kubenswrapper[4733]: I0202 15:16:15.923599 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:15 crc kubenswrapper[4733]: I0202 15:16:15.923657 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:15 crc kubenswrapper[4733]: I0202 15:16:15.923674 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:15 crc kubenswrapper[4733]: I0202 15:16:15.923697 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:15 crc kubenswrapper[4733]: I0202 15:16:15.923715 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:15Z","lastTransitionTime":"2026-02-02T15:16:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:16 crc kubenswrapper[4733]: I0202 15:16:16.027061 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:16 crc kubenswrapper[4733]: I0202 15:16:16.027100 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:16 crc kubenswrapper[4733]: I0202 15:16:16.027140 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:16 crc kubenswrapper[4733]: I0202 15:16:16.027226 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:16 crc kubenswrapper[4733]: I0202 15:16:16.027245 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:16Z","lastTransitionTime":"2026-02-02T15:16:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:16 crc kubenswrapper[4733]: I0202 15:16:16.130469 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:16 crc kubenswrapper[4733]: I0202 15:16:16.130528 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:16 crc kubenswrapper[4733]: I0202 15:16:16.130552 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:16 crc kubenswrapper[4733]: I0202 15:16:16.130578 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:16 crc kubenswrapper[4733]: I0202 15:16:16.130600 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:16Z","lastTransitionTime":"2026-02-02T15:16:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:16 crc kubenswrapper[4733]: I0202 15:16:16.227691 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 11:11:24.943789377 +0000 UTC Feb 02 15:16:16 crc kubenswrapper[4733]: I0202 15:16:16.233651 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:16 crc kubenswrapper[4733]: I0202 15:16:16.233712 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:16 crc kubenswrapper[4733]: I0202 15:16:16.233730 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:16 crc kubenswrapper[4733]: I0202 15:16:16.233755 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:16 crc kubenswrapper[4733]: I0202 15:16:16.233777 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:16Z","lastTransitionTime":"2026-02-02T15:16:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:16 crc kubenswrapper[4733]: I0202 15:16:16.337140 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:16 crc kubenswrapper[4733]: I0202 15:16:16.337238 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:16 crc kubenswrapper[4733]: I0202 15:16:16.337261 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:16 crc kubenswrapper[4733]: I0202 15:16:16.337295 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:16 crc kubenswrapper[4733]: I0202 15:16:16.337319 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:16Z","lastTransitionTime":"2026-02-02T15:16:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:16 crc kubenswrapper[4733]: I0202 15:16:16.440209 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:16 crc kubenswrapper[4733]: I0202 15:16:16.440272 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:16 crc kubenswrapper[4733]: I0202 15:16:16.440294 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:16 crc kubenswrapper[4733]: I0202 15:16:16.440324 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:16 crc kubenswrapper[4733]: I0202 15:16:16.440347 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:16Z","lastTransitionTime":"2026-02-02T15:16:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:16 crc kubenswrapper[4733]: I0202 15:16:16.543242 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:16 crc kubenswrapper[4733]: I0202 15:16:16.543298 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:16 crc kubenswrapper[4733]: I0202 15:16:16.543322 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:16 crc kubenswrapper[4733]: I0202 15:16:16.543354 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:16 crc kubenswrapper[4733]: I0202 15:16:16.543377 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:16Z","lastTransitionTime":"2026-02-02T15:16:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:16 crc kubenswrapper[4733]: I0202 15:16:16.646381 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:16 crc kubenswrapper[4733]: I0202 15:16:16.646440 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:16 crc kubenswrapper[4733]: I0202 15:16:16.646456 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:16 crc kubenswrapper[4733]: I0202 15:16:16.646479 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:16 crc kubenswrapper[4733]: I0202 15:16:16.646496 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:16Z","lastTransitionTime":"2026-02-02T15:16:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:16 crc kubenswrapper[4733]: I0202 15:16:16.749534 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:16 crc kubenswrapper[4733]: I0202 15:16:16.749596 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:16 crc kubenswrapper[4733]: I0202 15:16:16.749614 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:16 crc kubenswrapper[4733]: I0202 15:16:16.749638 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:16 crc kubenswrapper[4733]: I0202 15:16:16.749656 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:16Z","lastTransitionTime":"2026-02-02T15:16:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:16 crc kubenswrapper[4733]: I0202 15:16:16.852559 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:16 crc kubenswrapper[4733]: I0202 15:16:16.852614 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:16 crc kubenswrapper[4733]: I0202 15:16:16.852631 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:16 crc kubenswrapper[4733]: I0202 15:16:16.852660 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:16 crc kubenswrapper[4733]: I0202 15:16:16.852685 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:16Z","lastTransitionTime":"2026-02-02T15:16:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:16 crc kubenswrapper[4733]: I0202 15:16:16.955722 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:16 crc kubenswrapper[4733]: I0202 15:16:16.955795 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:16 crc kubenswrapper[4733]: I0202 15:16:16.955814 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:16 crc kubenswrapper[4733]: I0202 15:16:16.955837 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:16 crc kubenswrapper[4733]: I0202 15:16:16.955854 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:16Z","lastTransitionTime":"2026-02-02T15:16:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.059495 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.059547 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.059566 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.059591 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.059610 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:17Z","lastTransitionTime":"2026-02-02T15:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.162651 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.162795 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.162823 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.162851 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.162871 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:17Z","lastTransitionTime":"2026-02-02T15:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.229691 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 08:41:10.06635607 +0000 UTC Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.254290 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.254471 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:16:17 crc kubenswrapper[4733]: E0202 15:16:17.254672 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.254739 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.254829 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:16:17 crc kubenswrapper[4733]: E0202 15:16:17.255013 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:16:17 crc kubenswrapper[4733]: E0202 15:16:17.255088 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:16:17 crc kubenswrapper[4733]: E0202 15:16:17.255272 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.265657 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.265721 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.265744 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.265772 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.267251 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:17Z","lastTransitionTime":"2026-02-02T15:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.277021 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"402a6dc5-6bc7-4d7c-a909-9bb4ca48365d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d5b032497b592363a03f89cb31ba1b551f479004072dddffce540c41190ab81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a3fd02d55a10647428a05a9f53a1429a9827059c523d5deab3c4581c8f1b933\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11600ada13fbc7db1831de385202bc1f041efb656cc58aa49b0f71fe902db9bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fd39f5a05dec7a756dfd519f5df3ef7c77bbb5b8633c5af18a09ff8029cb079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:17Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.297857 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab21f7f0-7a5d-4eed-bd82-6c9d2e697eda\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ae3c8e412002f52fe150522d9bacb6415308ea41825b90d9d89010a7eca8764\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73fee4346d0af07556e8a1ea336818acea1f39aa8f703056b5954c7fac72aee1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdff2efc8ee6095385be74fbe1bf30926e526639dcc00b39fca8d5c9a0145d3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://742ee25d4d54068f173176593cafc678ec8502b53656fe9c969177f803836e29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://742ee25d4d54068f173176593cafc678ec8502b53656fe9c969177f803836e29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:17Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.318653 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:17Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.339493 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b38b8410bea95465ee58803c4160039bf7c082ca47893680da7469418f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27003b336112760daa2079f4819641850c27fbc8415b63b5094c86195291aca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:17Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.365836 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jh6vk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76a05d6-d9c7-4183-ad90-b3131f225841\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f2659e2a914d3191daddbfb5ad109e487f03dabb559d324244e127add7d41d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jh6vk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:17Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.369889 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.369937 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.369953 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.369978 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.369996 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:17Z","lastTransitionTime":"2026-02-02T15:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.385952 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83d0e810-ea5c-47aa-947b-7865b045bf94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6144eb32f88bc106f7724e48133b568a137c4f3fbbfe30c7fcef405d3ab735a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51318ad71b42f4cc8d376aa50e44580e68da89d83d80defb996329a2ee5c2ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zgsv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:17Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.401832 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7r9fd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4dfd44b-b363-4162-9f43-9a8dc4a3b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7r9fd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:17Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.426074 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7082da2-f2e7-4b73-ba35-504e5cd4c920\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd383d69b9b27d52fa8672ff6e0294a74876d676973bdc0ca95125fef5f2d085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037e379a109ce230098e3ca223fd1dea3149de3ebe601f56e2bb295f3093a3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6fdd19dac506f2d5d7eedc44ec2cda92a9381b73e91a45d949e4fcde4dfc283\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13965f5abcfb0a36c33310518955f192673e630a1aaba94b49400026de5bc2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://992f4198192cbe4b1c89d575aa4a20c29c1115fd5eaa8986c6ed70c1d6b143c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0202 15:15:31.010033 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 15:15:31.011243 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3543528698/tls.crt::/tmp/serving-cert-3543528698/tls.key\\\\\\\"\\\\nI0202 15:15:36.416658 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 15:15:36.428087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 15:15:36.428134 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 15:15:36.428243 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 15:15:36.428280 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 15:15:36.439678 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 15:15:36.439710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439717 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439722 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 15:15:36.439726 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 15:15:36.439729 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 15:15:36.439732 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 15:15:36.439753 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 15:15:36.450716 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://226f5949820d4f498dbf6fa251c6edf1c6356900bd946e6afae2c29042b1c835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:17Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.448749 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da08f9aecc070b31df2ddc1be7c6eb2ad8ef68a931e57ca2e5c5302f4e75751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:17Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.469634 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:17Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.473024 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.473064 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.473080 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.473133 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.473156 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:17Z","lastTransitionTime":"2026-02-02T15:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.483366 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.483448 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.483473 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.483505 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.483529 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:17Z","lastTransitionTime":"2026-02-02T15:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.488604 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca57dabf1d1ee5c6fc29291b1be687ca7e97c72650084b02067057fccac41a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:17Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:17 crc kubenswrapper[4733]: E0202 15:16:17.503453 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd742269-802e-45ec-a88e-17aee17c574e\\\",\\\"systemUUID\\\":\\\"91b64b75-3cd9-4fb9-93f0-dbd3bcdcaec4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:17Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.508075 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:17Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.509520 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.509570 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.509586 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.509649 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.509667 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:17Z","lastTransitionTime":"2026-02-02T15:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.524329 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5xgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56681968-f264-440e-aa12-d4f81879efe9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e250891b1065a38a448aa6abd300b37543e819e87f006b03cc37983cb643982a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdjjl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5xgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:17Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:17 crc kubenswrapper[4733]: E0202 15:16:17.527814 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd742269-802e-45ec-a88e-17aee17c574e\\\",\\\"systemUUID\\\":\\\"91b64b75-3cd9-4fb9-93f0-dbd3bcdcaec4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:17Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.533928 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.534028 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.534045 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.534067 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.534086 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:17Z","lastTransitionTime":"2026-02-02T15:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:17 crc kubenswrapper[4733]: E0202 15:16:17.552485 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd742269-802e-45ec-a88e-17aee17c574e\\\",\\\"systemUUID\\\":\\\"91b64b75-3cd9-4fb9-93f0-dbd3bcdcaec4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:17Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.557718 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.558005 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.558240 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.558458 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.558724 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:17Z","lastTransitionTime":"2026-02-02T15:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.560118 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8206eca4-7d8b-4b59-a4c1-24631dd8cd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41d4a7bdc21f00b5039385d7227d8cb0f4ea3051d3b5065c71a49cb0bfad26a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c11b1859bf3bc0ec4d5e6a58d9becc1ac6333a1ed9820d83d634f1e93221c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c08b087d85be9bb83007c7e38a5de9182a5091716fb8a2ca5ded1b8191bebc40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fd4969f0f16c9aa44208d2bf578f77d9d50914a8cb1764507175aa2eb58e733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59175c8e5a7f3536168a69fd944bf3d536bcd7281eee96002343b7836a54401b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:17Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:17 crc kubenswrapper[4733]: E0202 15:16:17.582481 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd742269-802e-45ec-a88e-17aee17c574e\\\",\\\"systemUUID\\\":\\\"91b64b75-3cd9-4fb9-93f0-dbd3bcdcaec4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:17Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.586129 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nr6ps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e43cd902-791a-4042-bc08-cf03e075be32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2a89a5d1fc43696f33d313aee5dca316a249fdef6a23ff655e31f14cc449cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07e56782051c93b2a320aa539a42764b97647237d228bed2b6af48cee987c60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07e56782051c93b2a320aa539a42764b97647237d228bed2b6af48cee987c60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce850ac7bd4c2444b65d87232114d55b78d11c6f973537c615900092a45d55e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce850ac7bd4c2444b65d87232114d55b78d11c6f973537c615900092a45d55e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nr6ps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:17Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.588189 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.588249 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.588269 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.588291 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.588310 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:17Z","lastTransitionTime":"2026-02-02T15:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:17 crc kubenswrapper[4733]: E0202 15:16:17.609293 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd742269-802e-45ec-a88e-17aee17c574e\\\",\\\"systemUUID\\\":\\\"91b64b75-3cd9-4fb9-93f0-dbd3bcdcaec4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:17Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:17 crc kubenswrapper[4733]: E0202 15:16:17.609582 4733 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.611964 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.612036 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.612062 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.612094 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.612118 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:17Z","lastTransitionTime":"2026-02-02T15:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.618401 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5a81ada-e393-499a-aac6-7b40349958bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4acf2398e2e19bd54f482cb2273c54738e61627187860bb0b6daadddc2565b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831d85055203f153381eb51d98af0cd2e4c4a0309f837545ab3d855a680133e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1380122d1694c1286f1102e6396c6504b91f63c6eb157689dfc0ea7188c17f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5835cb53385359e9c58eaa5d4da2569c313e0fc265ad79a6c124e8cecb1a734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd6dc4bf5b9fd31bfbee6d9340fe981c0c5fca03a79cef9316340ff5a8b3d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db1db75e392ebea8b387b85cabc4276d553bf2c06694cf7f294a383909f32826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa356e9d1cf1236a6d325e762401b42848ed4c48bc4c538a5844e09de54d16c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa356e9d1cf1236a6d325e762401b42848ed4c48bc4c538a5844e09de54d16c1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T15:16:02Z\\\",\\\"message\\\":\\\"\\\\nI0202 15:16:02.301761 6382 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-5xgpn in node crc\\\\nI0202 15:16:02.301771 6382 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-8c597\\\\nI0202 15:16:02.301779 6382 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-5xgpn after 0 failed attempt(s)\\\\nI0202 15:16:02.301784 6382 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-8c597 in node crc\\\\nI0202 15:16:02.301793 6382 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-5xgpn\\\\nI0202 15:16:02.301794 6382 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-8c597 after 0 failed attempt(s)\\\\nI0202 15:16:02.301805 6382 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-8c597\\\\nI0202 15:16:02.301355 6382 services_controller.go:360] Finished syncing service metrics on namespace openshift-ingress-operator for network=default : 1.801637ms\\\\nI0202 15:16:02.301684 6382 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI0202 15:16:02.301833 6382 services_controller.go:356] Processing sync for service openshift-console-operator/metrics for network=default\\\\nI0202 15:16:02\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:16:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qxvkf_openshift-ovn-kubernetes(f5a81ada-e393-499a-aac6-7b40349958bb)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efaf2216e4c1750139f080a93b4b608737dab2883de5fddfd9a99248d517b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxvkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:17Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.637342 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8c597" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76cc68c3-9e90-45de-87b2-6ea1b68eadce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ac57831c1c880e64aaa7bedcf999133ee09fc5ecc5124b9d54ecedcc9a13c28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwr7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8c597\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:17Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.658593 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hbm26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45553dc5-54ea-4de9-914d-55b8fee477f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d112b1ea41c3d30a7cfa46d17db6aed0999eb0baaaf8a2f21c741d805a21451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d64j4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f76a1f655f871deb7e1364fcc5dbf6b564c4c386300bb49cb75f83518b64286e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d64j4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hbm26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:17Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.715589 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.715962 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.716099 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.716321 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.716481 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:17Z","lastTransitionTime":"2026-02-02T15:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.819482 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.819549 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.819568 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.819593 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.819611 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:17Z","lastTransitionTime":"2026-02-02T15:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.922802 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.922866 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.922882 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.922906 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:17 crc kubenswrapper[4733]: I0202 15:16:17.922924 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:17Z","lastTransitionTime":"2026-02-02T15:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:18 crc kubenswrapper[4733]: I0202 15:16:18.026976 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:18 crc kubenswrapper[4733]: I0202 15:16:18.027020 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:18 crc kubenswrapper[4733]: I0202 15:16:18.027036 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:18 crc kubenswrapper[4733]: I0202 15:16:18.027062 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:18 crc kubenswrapper[4733]: I0202 15:16:18.027078 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:18Z","lastTransitionTime":"2026-02-02T15:16:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:18 crc kubenswrapper[4733]: I0202 15:16:18.130488 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:18 crc kubenswrapper[4733]: I0202 15:16:18.130950 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:18 crc kubenswrapper[4733]: I0202 15:16:18.131203 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:18 crc kubenswrapper[4733]: I0202 15:16:18.131463 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:18 crc kubenswrapper[4733]: I0202 15:16:18.131701 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:18Z","lastTransitionTime":"2026-02-02T15:16:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:18 crc kubenswrapper[4733]: I0202 15:16:18.230897 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 17:27:16.2710686 +0000 UTC Feb 02 15:16:18 crc kubenswrapper[4733]: I0202 15:16:18.235417 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:18 crc kubenswrapper[4733]: I0202 15:16:18.235662 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:18 crc kubenswrapper[4733]: I0202 15:16:18.235815 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:18 crc kubenswrapper[4733]: I0202 15:16:18.235967 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:18 crc kubenswrapper[4733]: I0202 15:16:18.236116 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:18Z","lastTransitionTime":"2026-02-02T15:16:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:18 crc kubenswrapper[4733]: I0202 15:16:18.340245 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:18 crc kubenswrapper[4733]: I0202 15:16:18.340684 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:18 crc kubenswrapper[4733]: I0202 15:16:18.340709 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:18 crc kubenswrapper[4733]: I0202 15:16:18.340739 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:18 crc kubenswrapper[4733]: I0202 15:16:18.340765 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:18Z","lastTransitionTime":"2026-02-02T15:16:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:18 crc kubenswrapper[4733]: I0202 15:16:18.443818 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:18 crc kubenswrapper[4733]: I0202 15:16:18.443878 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:18 crc kubenswrapper[4733]: I0202 15:16:18.443899 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:18 crc kubenswrapper[4733]: I0202 15:16:18.443924 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:18 crc kubenswrapper[4733]: I0202 15:16:18.443942 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:18Z","lastTransitionTime":"2026-02-02T15:16:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:18 crc kubenswrapper[4733]: I0202 15:16:18.546356 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:18 crc kubenswrapper[4733]: I0202 15:16:18.546415 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:18 crc kubenswrapper[4733]: I0202 15:16:18.546439 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:18 crc kubenswrapper[4733]: I0202 15:16:18.546469 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:18 crc kubenswrapper[4733]: I0202 15:16:18.546489 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:18Z","lastTransitionTime":"2026-02-02T15:16:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:18 crc kubenswrapper[4733]: I0202 15:16:18.649324 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:18 crc kubenswrapper[4733]: I0202 15:16:18.649671 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:18 crc kubenswrapper[4733]: I0202 15:16:18.649826 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:18 crc kubenswrapper[4733]: I0202 15:16:18.649967 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:18 crc kubenswrapper[4733]: I0202 15:16:18.650103 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:18Z","lastTransitionTime":"2026-02-02T15:16:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:18 crc kubenswrapper[4733]: I0202 15:16:18.753225 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:18 crc kubenswrapper[4733]: I0202 15:16:18.753281 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:18 crc kubenswrapper[4733]: I0202 15:16:18.753304 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:18 crc kubenswrapper[4733]: I0202 15:16:18.753331 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:18 crc kubenswrapper[4733]: I0202 15:16:18.753348 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:18Z","lastTransitionTime":"2026-02-02T15:16:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:18 crc kubenswrapper[4733]: I0202 15:16:18.856779 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:18 crc kubenswrapper[4733]: I0202 15:16:18.856857 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:18 crc kubenswrapper[4733]: I0202 15:16:18.856879 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:18 crc kubenswrapper[4733]: I0202 15:16:18.856903 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:18 crc kubenswrapper[4733]: I0202 15:16:18.856921 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:18Z","lastTransitionTime":"2026-02-02T15:16:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:18 crc kubenswrapper[4733]: I0202 15:16:18.960069 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:18 crc kubenswrapper[4733]: I0202 15:16:18.960232 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:18 crc kubenswrapper[4733]: I0202 15:16:18.960268 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:18 crc kubenswrapper[4733]: I0202 15:16:18.960297 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:18 crc kubenswrapper[4733]: I0202 15:16:18.960321 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:18Z","lastTransitionTime":"2026-02-02T15:16:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.064385 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.064432 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.064448 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.064470 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.064487 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:19Z","lastTransitionTime":"2026-02-02T15:16:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.167871 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.168015 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.168033 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.168059 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.168101 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:19Z","lastTransitionTime":"2026-02-02T15:16:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.231915 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-08 13:19:36.347233162 +0000 UTC Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.254418 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.254578 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:16:19 crc kubenswrapper[4733]: E0202 15:16:19.254819 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.254887 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.254854 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:16:19 crc kubenswrapper[4733]: E0202 15:16:19.255099 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:16:19 crc kubenswrapper[4733]: E0202 15:16:19.255324 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:16:19 crc kubenswrapper[4733]: E0202 15:16:19.255437 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.271450 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.271538 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.271560 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.271582 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.271599 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:19Z","lastTransitionTime":"2026-02-02T15:16:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.374123 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.374226 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.374246 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.374269 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.374287 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:19Z","lastTransitionTime":"2026-02-02T15:16:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.478210 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.478281 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.478299 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.478324 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.478343 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:19Z","lastTransitionTime":"2026-02-02T15:16:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.580916 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.580981 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.580998 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.581021 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.581038 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:19Z","lastTransitionTime":"2026-02-02T15:16:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.683995 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.684067 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.684095 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.684125 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.684151 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:19Z","lastTransitionTime":"2026-02-02T15:16:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.787717 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.787779 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.787801 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.787831 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.787854 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:19Z","lastTransitionTime":"2026-02-02T15:16:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.890379 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.890422 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.890434 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.890450 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.890463 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:19Z","lastTransitionTime":"2026-02-02T15:16:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.993918 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.993962 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.993975 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.993990 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:19 crc kubenswrapper[4733]: I0202 15:16:19.994002 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:19Z","lastTransitionTime":"2026-02-02T15:16:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:20 crc kubenswrapper[4733]: I0202 15:16:20.096674 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:20 crc kubenswrapper[4733]: I0202 15:16:20.096733 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:20 crc kubenswrapper[4733]: I0202 15:16:20.096749 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:20 crc kubenswrapper[4733]: I0202 15:16:20.096771 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:20 crc kubenswrapper[4733]: I0202 15:16:20.096788 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:20Z","lastTransitionTime":"2026-02-02T15:16:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:20 crc kubenswrapper[4733]: I0202 15:16:20.200045 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:20 crc kubenswrapper[4733]: I0202 15:16:20.200099 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:20 crc kubenswrapper[4733]: I0202 15:16:20.200123 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:20 crc kubenswrapper[4733]: I0202 15:16:20.200153 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:20 crc kubenswrapper[4733]: I0202 15:16:20.200209 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:20Z","lastTransitionTime":"2026-02-02T15:16:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:20 crc kubenswrapper[4733]: I0202 15:16:20.232644 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 23:07:19.344381493 +0000 UTC Feb 02 15:16:20 crc kubenswrapper[4733]: I0202 15:16:20.302658 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:20 crc kubenswrapper[4733]: I0202 15:16:20.302718 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:20 crc kubenswrapper[4733]: I0202 15:16:20.302739 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:20 crc kubenswrapper[4733]: I0202 15:16:20.302764 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:20 crc kubenswrapper[4733]: I0202 15:16:20.302785 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:20Z","lastTransitionTime":"2026-02-02T15:16:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:20 crc kubenswrapper[4733]: I0202 15:16:20.408654 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:20 crc kubenswrapper[4733]: I0202 15:16:20.408731 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:20 crc kubenswrapper[4733]: I0202 15:16:20.408751 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:20 crc kubenswrapper[4733]: I0202 15:16:20.408778 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:20 crc kubenswrapper[4733]: I0202 15:16:20.408802 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:20Z","lastTransitionTime":"2026-02-02T15:16:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:20 crc kubenswrapper[4733]: I0202 15:16:20.511671 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:20 crc kubenswrapper[4733]: I0202 15:16:20.511731 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:20 crc kubenswrapper[4733]: I0202 15:16:20.511748 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:20 crc kubenswrapper[4733]: I0202 15:16:20.511773 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:20 crc kubenswrapper[4733]: I0202 15:16:20.511791 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:20Z","lastTransitionTime":"2026-02-02T15:16:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:20 crc kubenswrapper[4733]: I0202 15:16:20.614754 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:20 crc kubenswrapper[4733]: I0202 15:16:20.614815 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:20 crc kubenswrapper[4733]: I0202 15:16:20.614831 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:20 crc kubenswrapper[4733]: I0202 15:16:20.614855 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:20 crc kubenswrapper[4733]: I0202 15:16:20.614874 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:20Z","lastTransitionTime":"2026-02-02T15:16:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:20 crc kubenswrapper[4733]: I0202 15:16:20.718103 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:20 crc kubenswrapper[4733]: I0202 15:16:20.718235 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:20 crc kubenswrapper[4733]: I0202 15:16:20.718256 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:20 crc kubenswrapper[4733]: I0202 15:16:20.718280 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:20 crc kubenswrapper[4733]: I0202 15:16:20.718297 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:20Z","lastTransitionTime":"2026-02-02T15:16:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:20 crc kubenswrapper[4733]: I0202 15:16:20.821776 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:20 crc kubenswrapper[4733]: I0202 15:16:20.822137 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:20 crc kubenswrapper[4733]: I0202 15:16:20.822322 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:20 crc kubenswrapper[4733]: I0202 15:16:20.822478 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:20 crc kubenswrapper[4733]: I0202 15:16:20.822619 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:20Z","lastTransitionTime":"2026-02-02T15:16:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:20 crc kubenswrapper[4733]: I0202 15:16:20.925833 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:20 crc kubenswrapper[4733]: I0202 15:16:20.926241 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:20 crc kubenswrapper[4733]: I0202 15:16:20.926447 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:20 crc kubenswrapper[4733]: I0202 15:16:20.926603 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:20 crc kubenswrapper[4733]: I0202 15:16:20.926788 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:20Z","lastTransitionTime":"2026-02-02T15:16:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.030410 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.030439 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.030468 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.030482 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.030490 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:21Z","lastTransitionTime":"2026-02-02T15:16:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.133311 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.133348 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.133360 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.133375 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.133388 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:21Z","lastTransitionTime":"2026-02-02T15:16:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.232985 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 05:57:21.161813665 +0000 UTC Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.235869 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.235902 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.235912 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.235925 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.235934 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:21Z","lastTransitionTime":"2026-02-02T15:16:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.254469 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.254514 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.254581 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:16:21 crc kubenswrapper[4733]: E0202 15:16:21.254586 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.254598 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:16:21 crc kubenswrapper[4733]: E0202 15:16:21.254666 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:16:21 crc kubenswrapper[4733]: E0202 15:16:21.254753 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:16:21 crc kubenswrapper[4733]: E0202 15:16:21.254818 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.338728 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.338773 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.338783 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.338798 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.338807 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:21Z","lastTransitionTime":"2026-02-02T15:16:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.441580 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.441634 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.441653 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.441675 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.441692 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:21Z","lastTransitionTime":"2026-02-02T15:16:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.544136 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.544220 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.544237 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.544260 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.544277 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:21Z","lastTransitionTime":"2026-02-02T15:16:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.646787 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.646833 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.646850 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.646872 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.646889 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:21Z","lastTransitionTime":"2026-02-02T15:16:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.750738 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.750784 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.750800 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.750822 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.750838 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:21Z","lastTransitionTime":"2026-02-02T15:16:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.854132 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.854235 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.854261 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.854291 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.854315 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:21Z","lastTransitionTime":"2026-02-02T15:16:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.957665 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.957725 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.957742 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.957769 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:21 crc kubenswrapper[4733]: I0202 15:16:21.957785 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:21Z","lastTransitionTime":"2026-02-02T15:16:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:22 crc kubenswrapper[4733]: I0202 15:16:22.060558 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:22 crc kubenswrapper[4733]: I0202 15:16:22.060616 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:22 crc kubenswrapper[4733]: I0202 15:16:22.060633 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:22 crc kubenswrapper[4733]: I0202 15:16:22.060657 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:22 crc kubenswrapper[4733]: I0202 15:16:22.060679 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:22Z","lastTransitionTime":"2026-02-02T15:16:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:22 crc kubenswrapper[4733]: I0202 15:16:22.163220 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:22 crc kubenswrapper[4733]: I0202 15:16:22.163280 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:22 crc kubenswrapper[4733]: I0202 15:16:22.163288 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:22 crc kubenswrapper[4733]: I0202 15:16:22.163307 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:22 crc kubenswrapper[4733]: I0202 15:16:22.163317 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:22Z","lastTransitionTime":"2026-02-02T15:16:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:22 crc kubenswrapper[4733]: I0202 15:16:22.234669 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 16:09:42.731317045 +0000 UTC Feb 02 15:16:22 crc kubenswrapper[4733]: I0202 15:16:22.265781 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:22 crc kubenswrapper[4733]: I0202 15:16:22.265819 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:22 crc kubenswrapper[4733]: I0202 15:16:22.265827 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:22 crc kubenswrapper[4733]: I0202 15:16:22.265842 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:22 crc kubenswrapper[4733]: I0202 15:16:22.265851 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:22Z","lastTransitionTime":"2026-02-02T15:16:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:22 crc kubenswrapper[4733]: I0202 15:16:22.368768 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:22 crc kubenswrapper[4733]: I0202 15:16:22.368833 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:22 crc kubenswrapper[4733]: I0202 15:16:22.368857 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:22 crc kubenswrapper[4733]: I0202 15:16:22.368889 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:22 crc kubenswrapper[4733]: I0202 15:16:22.368913 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:22Z","lastTransitionTime":"2026-02-02T15:16:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:22 crc kubenswrapper[4733]: I0202 15:16:22.471745 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:22 crc kubenswrapper[4733]: I0202 15:16:22.471807 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:22 crc kubenswrapper[4733]: I0202 15:16:22.471823 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:22 crc kubenswrapper[4733]: I0202 15:16:22.471849 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:22 crc kubenswrapper[4733]: I0202 15:16:22.471866 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:22Z","lastTransitionTime":"2026-02-02T15:16:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:22 crc kubenswrapper[4733]: I0202 15:16:22.575854 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:22 crc kubenswrapper[4733]: I0202 15:16:22.575895 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:22 crc kubenswrapper[4733]: I0202 15:16:22.575906 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:22 crc kubenswrapper[4733]: I0202 15:16:22.575922 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:22 crc kubenswrapper[4733]: I0202 15:16:22.575933 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:22Z","lastTransitionTime":"2026-02-02T15:16:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:22 crc kubenswrapper[4733]: I0202 15:16:22.678672 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:22 crc kubenswrapper[4733]: I0202 15:16:22.678733 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:22 crc kubenswrapper[4733]: I0202 15:16:22.678752 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:22 crc kubenswrapper[4733]: I0202 15:16:22.678778 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:22 crc kubenswrapper[4733]: I0202 15:16:22.678795 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:22Z","lastTransitionTime":"2026-02-02T15:16:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:22 crc kubenswrapper[4733]: I0202 15:16:22.781449 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:22 crc kubenswrapper[4733]: I0202 15:16:22.781512 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:22 crc kubenswrapper[4733]: I0202 15:16:22.781529 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:22 crc kubenswrapper[4733]: I0202 15:16:22.781555 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:22 crc kubenswrapper[4733]: I0202 15:16:22.781574 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:22Z","lastTransitionTime":"2026-02-02T15:16:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:22 crc kubenswrapper[4733]: I0202 15:16:22.884637 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:22 crc kubenswrapper[4733]: I0202 15:16:22.884679 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:22 crc kubenswrapper[4733]: I0202 15:16:22.884688 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:22 crc kubenswrapper[4733]: I0202 15:16:22.884705 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:22 crc kubenswrapper[4733]: I0202 15:16:22.884714 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:22Z","lastTransitionTime":"2026-02-02T15:16:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:22 crc kubenswrapper[4733]: I0202 15:16:22.986629 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:22 crc kubenswrapper[4733]: I0202 15:16:22.986665 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:22 crc kubenswrapper[4733]: I0202 15:16:22.986673 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:22 crc kubenswrapper[4733]: I0202 15:16:22.986688 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:22 crc kubenswrapper[4733]: I0202 15:16:22.986698 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:22Z","lastTransitionTime":"2026-02-02T15:16:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:23 crc kubenswrapper[4733]: I0202 15:16:23.089138 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:23 crc kubenswrapper[4733]: I0202 15:16:23.089229 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:23 crc kubenswrapper[4733]: I0202 15:16:23.089248 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:23 crc kubenswrapper[4733]: I0202 15:16:23.089272 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:23 crc kubenswrapper[4733]: I0202 15:16:23.089289 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:23Z","lastTransitionTime":"2026-02-02T15:16:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:23 crc kubenswrapper[4733]: I0202 15:16:23.191959 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:23 crc kubenswrapper[4733]: I0202 15:16:23.192057 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:23 crc kubenswrapper[4733]: I0202 15:16:23.192077 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:23 crc kubenswrapper[4733]: I0202 15:16:23.192153 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:23 crc kubenswrapper[4733]: I0202 15:16:23.192231 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:23Z","lastTransitionTime":"2026-02-02T15:16:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:23 crc kubenswrapper[4733]: I0202 15:16:23.235734 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-26 08:44:33.141246378 +0000 UTC Feb 02 15:16:23 crc kubenswrapper[4733]: I0202 15:16:23.254086 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:16:23 crc kubenswrapper[4733]: I0202 15:16:23.254222 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:16:23 crc kubenswrapper[4733]: E0202 15:16:23.254313 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:16:23 crc kubenswrapper[4733]: I0202 15:16:23.254405 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:16:23 crc kubenswrapper[4733]: E0202 15:16:23.254590 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:16:23 crc kubenswrapper[4733]: I0202 15:16:23.254664 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:16:23 crc kubenswrapper[4733]: E0202 15:16:23.254730 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:16:23 crc kubenswrapper[4733]: E0202 15:16:23.254872 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:16:23 crc kubenswrapper[4733]: I0202 15:16:23.295293 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:23 crc kubenswrapper[4733]: I0202 15:16:23.295334 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:23 crc kubenswrapper[4733]: I0202 15:16:23.295345 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:23 crc kubenswrapper[4733]: I0202 15:16:23.295364 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:23 crc kubenswrapper[4733]: I0202 15:16:23.295375 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:23Z","lastTransitionTime":"2026-02-02T15:16:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:23 crc kubenswrapper[4733]: I0202 15:16:23.398346 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:23 crc kubenswrapper[4733]: I0202 15:16:23.398392 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:23 crc kubenswrapper[4733]: I0202 15:16:23.398409 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:23 crc kubenswrapper[4733]: I0202 15:16:23.398432 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:23 crc kubenswrapper[4733]: I0202 15:16:23.398448 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:23Z","lastTransitionTime":"2026-02-02T15:16:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:23 crc kubenswrapper[4733]: I0202 15:16:23.501961 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:23 crc kubenswrapper[4733]: I0202 15:16:23.501997 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:23 crc kubenswrapper[4733]: I0202 15:16:23.502008 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:23 crc kubenswrapper[4733]: I0202 15:16:23.502021 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:23 crc kubenswrapper[4733]: I0202 15:16:23.502031 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:23Z","lastTransitionTime":"2026-02-02T15:16:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:23 crc kubenswrapper[4733]: I0202 15:16:23.603800 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:23 crc kubenswrapper[4733]: I0202 15:16:23.603838 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:23 crc kubenswrapper[4733]: I0202 15:16:23.603847 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:23 crc kubenswrapper[4733]: I0202 15:16:23.603859 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:23 crc kubenswrapper[4733]: I0202 15:16:23.603867 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:23Z","lastTransitionTime":"2026-02-02T15:16:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:23 crc kubenswrapper[4733]: I0202 15:16:23.705683 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:23 crc kubenswrapper[4733]: I0202 15:16:23.705757 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:23 crc kubenswrapper[4733]: I0202 15:16:23.705776 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:23 crc kubenswrapper[4733]: I0202 15:16:23.705799 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:23 crc kubenswrapper[4733]: I0202 15:16:23.705815 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:23Z","lastTransitionTime":"2026-02-02T15:16:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:23 crc kubenswrapper[4733]: I0202 15:16:23.808903 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:23 crc kubenswrapper[4733]: I0202 15:16:23.808947 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:23 crc kubenswrapper[4733]: I0202 15:16:23.808963 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:23 crc kubenswrapper[4733]: I0202 15:16:23.808985 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:23 crc kubenswrapper[4733]: I0202 15:16:23.809002 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:23Z","lastTransitionTime":"2026-02-02T15:16:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:23 crc kubenswrapper[4733]: I0202 15:16:23.911228 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:23 crc kubenswrapper[4733]: I0202 15:16:23.911258 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:23 crc kubenswrapper[4733]: I0202 15:16:23.911270 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:23 crc kubenswrapper[4733]: I0202 15:16:23.911287 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:23 crc kubenswrapper[4733]: I0202 15:16:23.911299 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:23Z","lastTransitionTime":"2026-02-02T15:16:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:24 crc kubenswrapper[4733]: I0202 15:16:24.013562 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:24 crc kubenswrapper[4733]: I0202 15:16:24.013818 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:24 crc kubenswrapper[4733]: I0202 15:16:24.013911 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:24 crc kubenswrapper[4733]: I0202 15:16:24.013991 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:24 crc kubenswrapper[4733]: I0202 15:16:24.014077 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:24Z","lastTransitionTime":"2026-02-02T15:16:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:24 crc kubenswrapper[4733]: I0202 15:16:24.116044 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:24 crc kubenswrapper[4733]: I0202 15:16:24.116073 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:24 crc kubenswrapper[4733]: I0202 15:16:24.116081 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:24 crc kubenswrapper[4733]: I0202 15:16:24.116093 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:24 crc kubenswrapper[4733]: I0202 15:16:24.116102 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:24Z","lastTransitionTime":"2026-02-02T15:16:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:24 crc kubenswrapper[4733]: I0202 15:16:24.219115 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:24 crc kubenswrapper[4733]: I0202 15:16:24.219269 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:24 crc kubenswrapper[4733]: I0202 15:16:24.219299 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:24 crc kubenswrapper[4733]: I0202 15:16:24.219325 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:24 crc kubenswrapper[4733]: I0202 15:16:24.219559 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:24Z","lastTransitionTime":"2026-02-02T15:16:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:24 crc kubenswrapper[4733]: I0202 15:16:24.236701 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 20:09:03.413774812 +0000 UTC Feb 02 15:16:24 crc kubenswrapper[4733]: I0202 15:16:24.322045 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:24 crc kubenswrapper[4733]: I0202 15:16:24.322132 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:24 crc kubenswrapper[4733]: I0202 15:16:24.322150 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:24 crc kubenswrapper[4733]: I0202 15:16:24.322199 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:24 crc kubenswrapper[4733]: I0202 15:16:24.322219 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:24Z","lastTransitionTime":"2026-02-02T15:16:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:24 crc kubenswrapper[4733]: I0202 15:16:24.424917 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:24 crc kubenswrapper[4733]: I0202 15:16:24.424982 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:24 crc kubenswrapper[4733]: I0202 15:16:24.424999 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:24 crc kubenswrapper[4733]: I0202 15:16:24.425025 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:24 crc kubenswrapper[4733]: I0202 15:16:24.425042 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:24Z","lastTransitionTime":"2026-02-02T15:16:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:24 crc kubenswrapper[4733]: I0202 15:16:24.527655 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:24 crc kubenswrapper[4733]: I0202 15:16:24.527699 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:24 crc kubenswrapper[4733]: I0202 15:16:24.527707 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:24 crc kubenswrapper[4733]: I0202 15:16:24.527721 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:24 crc kubenswrapper[4733]: I0202 15:16:24.527731 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:24Z","lastTransitionTime":"2026-02-02T15:16:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:24 crc kubenswrapper[4733]: I0202 15:16:24.630461 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:24 crc kubenswrapper[4733]: I0202 15:16:24.630511 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:24 crc kubenswrapper[4733]: I0202 15:16:24.630530 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:24 crc kubenswrapper[4733]: I0202 15:16:24.630555 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:24 crc kubenswrapper[4733]: I0202 15:16:24.630573 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:24Z","lastTransitionTime":"2026-02-02T15:16:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:24 crc kubenswrapper[4733]: I0202 15:16:24.733298 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:24 crc kubenswrapper[4733]: I0202 15:16:24.733342 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:24 crc kubenswrapper[4733]: I0202 15:16:24.733352 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:24 crc kubenswrapper[4733]: I0202 15:16:24.733370 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:24 crc kubenswrapper[4733]: I0202 15:16:24.733380 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:24Z","lastTransitionTime":"2026-02-02T15:16:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:24 crc kubenswrapper[4733]: I0202 15:16:24.836575 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:24 crc kubenswrapper[4733]: I0202 15:16:24.836606 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:24 crc kubenswrapper[4733]: I0202 15:16:24.836615 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:24 crc kubenswrapper[4733]: I0202 15:16:24.836629 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:24 crc kubenswrapper[4733]: I0202 15:16:24.836641 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:24Z","lastTransitionTime":"2026-02-02T15:16:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:24 crc kubenswrapper[4733]: I0202 15:16:24.939080 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:24 crc kubenswrapper[4733]: I0202 15:16:24.939191 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:24 crc kubenswrapper[4733]: I0202 15:16:24.939218 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:24 crc kubenswrapper[4733]: I0202 15:16:24.939249 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:24 crc kubenswrapper[4733]: I0202 15:16:24.939271 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:24Z","lastTransitionTime":"2026-02-02T15:16:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.041978 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.042032 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.042045 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.042061 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.042073 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:25Z","lastTransitionTime":"2026-02-02T15:16:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.143734 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.143797 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.143821 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.143851 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.143874 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:25Z","lastTransitionTime":"2026-02-02T15:16:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.236944 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-16 17:59:49.915673711 +0000 UTC Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.245377 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.245412 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.245422 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.245439 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.245448 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:25Z","lastTransitionTime":"2026-02-02T15:16:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.254707 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.254713 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.254733 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.254713 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:16:25 crc kubenswrapper[4733]: E0202 15:16:25.254795 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:16:25 crc kubenswrapper[4733]: E0202 15:16:25.254885 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:16:25 crc kubenswrapper[4733]: E0202 15:16:25.254917 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:16:25 crc kubenswrapper[4733]: E0202 15:16:25.254980 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.348074 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.348135 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.348152 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.348205 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.348222 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:25Z","lastTransitionTime":"2026-02-02T15:16:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.450433 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.450501 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.450517 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.450541 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.450559 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:25Z","lastTransitionTime":"2026-02-02T15:16:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.557566 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.557639 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.557660 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.557687 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.557712 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:25Z","lastTransitionTime":"2026-02-02T15:16:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.660352 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.660405 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.660422 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.660445 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.660462 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:25Z","lastTransitionTime":"2026-02-02T15:16:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.695016 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jh6vk_b76a05d6-d9c7-4183-ad90-b3131f225841/kube-multus/0.log" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.695105 4733 generic.go:334] "Generic (PLEG): container finished" podID="b76a05d6-d9c7-4183-ad90-b3131f225841" containerID="7f2659e2a914d3191daddbfb5ad109e487f03dabb559d324244e127add7d41d9" exitCode=1 Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.695159 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jh6vk" event={"ID":"b76a05d6-d9c7-4183-ad90-b3131f225841","Type":"ContainerDied","Data":"7f2659e2a914d3191daddbfb5ad109e487f03dabb559d324244e127add7d41d9"} Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.695799 4733 scope.go:117] "RemoveContainer" containerID="7f2659e2a914d3191daddbfb5ad109e487f03dabb559d324244e127add7d41d9" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.707968 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca57dabf1d1ee5c6fc29291b1be687ca7e97c72650084b02067057fccac41a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:25Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.725913 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:25Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.739593 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5xgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56681968-f264-440e-aa12-d4f81879efe9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e250891b1065a38a448aa6abd300b37543e819e87f006b03cc37983cb643982a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdjjl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5xgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:25Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.752436 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7r9fd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4dfd44b-b363-4162-9f43-9a8dc4a3b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7r9fd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:25Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.767909 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.767945 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.767953 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.767966 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.767975 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:25Z","lastTransitionTime":"2026-02-02T15:16:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.771252 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7082da2-f2e7-4b73-ba35-504e5cd4c920\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd383d69b9b27d52fa8672ff6e0294a74876d676973bdc0ca95125fef5f2d085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037e379a109ce230098e3ca223fd1dea3149de3ebe601f56e2bb295f3093a3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6fdd19dac506f2d5d7eedc44ec2cda92a9381b73e91a45d949e4fcde4dfc283\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13965f5abcfb0a36c33310518955f192673e630a1aaba94b49400026de5bc2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://992f4198192cbe4b1c89d575aa4a20c29c1115fd5eaa8986c6ed70c1d6b143c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0202 15:15:31.010033 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 15:15:31.011243 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3543528698/tls.crt::/tmp/serving-cert-3543528698/tls.key\\\\\\\"\\\\nI0202 15:15:36.416658 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 15:15:36.428087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 15:15:36.428134 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 15:15:36.428243 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 15:15:36.428280 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 15:15:36.439678 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 15:15:36.439710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439717 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439722 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 15:15:36.439726 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 15:15:36.439729 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 15:15:36.439732 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 15:15:36.439753 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 15:15:36.450716 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://226f5949820d4f498dbf6fa251c6edf1c6356900bd946e6afae2c29042b1c835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:25Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.785420 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da08f9aecc070b31df2ddc1be7c6eb2ad8ef68a931e57ca2e5c5302f4e75751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:25Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.799069 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:25Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.812101 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hbm26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45553dc5-54ea-4de9-914d-55b8fee477f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d112b1ea41c3d30a7cfa46d17db6aed0999eb0baaaf8a2f21c741d805a21451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d64j4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f76a1f655f871deb7e1364fcc5dbf6b564c4c386300bb49cb75f83518b64286e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d64j4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hbm26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:25Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.832105 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8206eca4-7d8b-4b59-a4c1-24631dd8cd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41d4a7bdc21f00b5039385d7227d8cb0f4ea3051d3b5065c71a49cb0bfad26a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c11b1859bf3bc0ec4d5e6a58d9becc1ac6333a1ed9820d83d634f1e93221c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c08b087d85be9bb83007c7e38a5de9182a5091716fb8a2ca5ded1b8191bebc40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fd4969f0f16c9aa44208d2bf578f77d9d50914a8cb1764507175aa2eb58e733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59175c8e5a7f3536168a69fd944bf3d536bcd7281eee96002343b7836a54401b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:25Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.850176 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nr6ps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e43cd902-791a-4042-bc08-cf03e075be32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2a89a5d1fc43696f33d313aee5dca316a249fdef6a23ff655e31f14cc449cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07e56782051c93b2a320aa539a42764b97647237d228bed2b6af48cee987c60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07e56782051c93b2a320aa539a42764b97647237d228bed2b6af48cee987c60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce850ac7bd4c2444b65d87232114d55b78d11c6f973537c615900092a45d55e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce850ac7bd4c2444b65d87232114d55b78d11c6f973537c615900092a45d55e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nr6ps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:25Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.870374 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5a81ada-e393-499a-aac6-7b40349958bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4acf2398e2e19bd54f482cb2273c54738e61627187860bb0b6daadddc2565b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831d85055203f153381eb51d98af0cd2e4c4a0309f837545ab3d855a680133e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1380122d1694c1286f1102e6396c6504b91f63c6eb157689dfc0ea7188c17f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5835cb53385359e9c58eaa5d4da2569c313e0fc265ad79a6c124e8cecb1a734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd6dc4bf5b9fd31bfbee6d9340fe981c0c5fca03a79cef9316340ff5a8b3d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db1db75e392ebea8b387b85cabc4276d553bf2c06694cf7f294a383909f32826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa356e9d1cf1236a6d325e762401b42848ed4c48bc4c538a5844e09de54d16c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa356e9d1cf1236a6d325e762401b42848ed4c48bc4c538a5844e09de54d16c1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T15:16:02Z\\\",\\\"message\\\":\\\"\\\\nI0202 15:16:02.301761 6382 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-5xgpn in node crc\\\\nI0202 15:16:02.301771 6382 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-8c597\\\\nI0202 15:16:02.301779 6382 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-5xgpn after 0 failed attempt(s)\\\\nI0202 15:16:02.301784 6382 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-8c597 in node crc\\\\nI0202 15:16:02.301793 6382 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-5xgpn\\\\nI0202 15:16:02.301794 6382 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-8c597 after 0 failed attempt(s)\\\\nI0202 15:16:02.301805 6382 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-8c597\\\\nI0202 15:16:02.301355 6382 services_controller.go:360] Finished syncing service metrics on namespace openshift-ingress-operator for network=default : 1.801637ms\\\\nI0202 15:16:02.301684 6382 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI0202 15:16:02.301833 6382 services_controller.go:356] Processing sync for service openshift-console-operator/metrics for network=default\\\\nI0202 15:16:02\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:16:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qxvkf_openshift-ovn-kubernetes(f5a81ada-e393-499a-aac6-7b40349958bb)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efaf2216e4c1750139f080a93b4b608737dab2883de5fddfd9a99248d517b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxvkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:25Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.870741 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.870760 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.870770 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.870786 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.870799 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:25Z","lastTransitionTime":"2026-02-02T15:16:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.889265 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8c597" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76cc68c3-9e90-45de-87b2-6ea1b68eadce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ac57831c1c880e64aaa7bedcf999133ee09fc5ecc5124b9d54ecedcc9a13c28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwr7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8c597\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:25Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.906479 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"402a6dc5-6bc7-4d7c-a909-9bb4ca48365d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d5b032497b592363a03f89cb31ba1b551f479004072dddffce540c41190ab81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a3fd02d55a10647428a05a9f53a1429a9827059c523d5deab3c4581c8f1b933\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11600ada13fbc7db1831de385202bc1f041efb656cc58aa49b0f71fe902db9bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fd39f5a05dec7a756dfd519f5df3ef7c77bbb5b8633c5af18a09ff8029cb079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:25Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.918682 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab21f7f0-7a5d-4eed-bd82-6c9d2e697eda\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ae3c8e412002f52fe150522d9bacb6415308ea41825b90d9d89010a7eca8764\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73fee4346d0af07556e8a1ea336818acea1f39aa8f703056b5954c7fac72aee1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdff2efc8ee6095385be74fbe1bf30926e526639dcc00b39fca8d5c9a0145d3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://742ee25d4d54068f173176593cafc678ec8502b53656fe9c969177f803836e29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://742ee25d4d54068f173176593cafc678ec8502b53656fe9c969177f803836e29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:25Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.932053 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:25Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.949443 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b38b8410bea95465ee58803c4160039bf7c082ca47893680da7469418f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27003b336112760daa2079f4819641850c27fbc8415b63b5094c86195291aca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:25Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.967041 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jh6vk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76a05d6-d9c7-4183-ad90-b3131f225841\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f2659e2a914d3191daddbfb5ad109e487f03dabb559d324244e127add7d41d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f2659e2a914d3191daddbfb5ad109e487f03dabb559d324244e127add7d41d9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T15:16:24Z\\\",\\\"message\\\":\\\"2026-02-02T15:15:39+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_bbf1a3b4-24d2-4558-bc15-4414386de7b4\\\\n2026-02-02T15:15:39+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_bbf1a3b4-24d2-4558-bc15-4414386de7b4 to /host/opt/cni/bin/\\\\n2026-02-02T15:15:39Z [verbose] multus-daemon started\\\\n2026-02-02T15:15:39Z [verbose] Readiness Indicator file check\\\\n2026-02-02T15:16:24Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jh6vk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:25Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.972803 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.972856 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.972872 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.972896 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.972913 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:25Z","lastTransitionTime":"2026-02-02T15:16:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:25 crc kubenswrapper[4733]: I0202 15:16:25.981909 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83d0e810-ea5c-47aa-947b-7865b045bf94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6144eb32f88bc106f7724e48133b568a137c4f3fbbfe30c7fcef405d3ab735a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51318ad71b42f4cc8d376aa50e44580e68da89d83d80defb996329a2ee5c2ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zgsv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:25Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.076263 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.076329 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.076346 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.076371 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.076390 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:26Z","lastTransitionTime":"2026-02-02T15:16:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.179265 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.179308 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.179318 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.179331 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.179341 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:26Z","lastTransitionTime":"2026-02-02T15:16:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.237778 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 19:34:51.604295664 +0000 UTC Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.282108 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.282206 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.282232 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.282261 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.282282 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:26Z","lastTransitionTime":"2026-02-02T15:16:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.385451 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.385498 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.385510 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.385526 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.385539 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:26Z","lastTransitionTime":"2026-02-02T15:16:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.489049 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.489108 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.489128 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.489152 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.489206 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:26Z","lastTransitionTime":"2026-02-02T15:16:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.592208 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.592287 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.592309 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.592339 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.592361 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:26Z","lastTransitionTime":"2026-02-02T15:16:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.694945 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.695020 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.695033 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.695049 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.695061 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:26Z","lastTransitionTime":"2026-02-02T15:16:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.700922 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jh6vk_b76a05d6-d9c7-4183-ad90-b3131f225841/kube-multus/0.log" Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.701037 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jh6vk" event={"ID":"b76a05d6-d9c7-4183-ad90-b3131f225841","Type":"ContainerStarted","Data":"f57aea989227fdbd7bb59c51fade50d1076929f6288d060808b51bd1bdd7df04"} Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.726116 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8206eca4-7d8b-4b59-a4c1-24631dd8cd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41d4a7bdc21f00b5039385d7227d8cb0f4ea3051d3b5065c71a49cb0bfad26a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c11b1859bf3bc0ec4d5e6a58d9becc1ac6333a1ed9820d83d634f1e93221c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c08b087d85be9bb83007c7e38a5de9182a5091716fb8a2ca5ded1b8191bebc40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fd4969f0f16c9aa44208d2bf578f77d9d50914a8cb1764507175aa2eb58e733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59175c8e5a7f3536168a69fd944bf3d536bcd7281eee96002343b7836a54401b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:26Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.747199 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nr6ps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e43cd902-791a-4042-bc08-cf03e075be32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2a89a5d1fc43696f33d313aee5dca316a249fdef6a23ff655e31f14cc449cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07e56782051c93b2a320aa539a42764b97647237d228bed2b6af48cee987c60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07e56782051c93b2a320aa539a42764b97647237d228bed2b6af48cee987c60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce850ac7bd4c2444b65d87232114d55b78d11c6f973537c615900092a45d55e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce850ac7bd4c2444b65d87232114d55b78d11c6f973537c615900092a45d55e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nr6ps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:26Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.771845 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5a81ada-e393-499a-aac6-7b40349958bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4acf2398e2e19bd54f482cb2273c54738e61627187860bb0b6daadddc2565b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831d85055203f153381eb51d98af0cd2e4c4a0309f837545ab3d855a680133e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1380122d1694c1286f1102e6396c6504b91f63c6eb157689dfc0ea7188c17f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5835cb53385359e9c58eaa5d4da2569c313e0fc265ad79a6c124e8cecb1a734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd6dc4bf5b9fd31bfbee6d9340fe981c0c5fca03a79cef9316340ff5a8b3d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db1db75e392ebea8b387b85cabc4276d553bf2c06694cf7f294a383909f32826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa356e9d1cf1236a6d325e762401b42848ed4c48bc4c538a5844e09de54d16c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa356e9d1cf1236a6d325e762401b42848ed4c48bc4c538a5844e09de54d16c1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T15:16:02Z\\\",\\\"message\\\":\\\"\\\\nI0202 15:16:02.301761 6382 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-5xgpn in node crc\\\\nI0202 15:16:02.301771 6382 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-8c597\\\\nI0202 15:16:02.301779 6382 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-5xgpn after 0 failed attempt(s)\\\\nI0202 15:16:02.301784 6382 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-8c597 in node crc\\\\nI0202 15:16:02.301793 6382 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-5xgpn\\\\nI0202 15:16:02.301794 6382 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-8c597 after 0 failed attempt(s)\\\\nI0202 15:16:02.301805 6382 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-8c597\\\\nI0202 15:16:02.301355 6382 services_controller.go:360] Finished syncing service metrics on namespace openshift-ingress-operator for network=default : 1.801637ms\\\\nI0202 15:16:02.301684 6382 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI0202 15:16:02.301833 6382 services_controller.go:356] Processing sync for service openshift-console-operator/metrics for network=default\\\\nI0202 15:16:02\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:16:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qxvkf_openshift-ovn-kubernetes(f5a81ada-e393-499a-aac6-7b40349958bb)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efaf2216e4c1750139f080a93b4b608737dab2883de5fddfd9a99248d517b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxvkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:26Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.784128 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8c597" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76cc68c3-9e90-45de-87b2-6ea1b68eadce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ac57831c1c880e64aaa7bedcf999133ee09fc5ecc5124b9d54ecedcc9a13c28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwr7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8c597\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:26Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.798011 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.798039 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.798049 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.798061 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.798071 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:26Z","lastTransitionTime":"2026-02-02T15:16:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.800323 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hbm26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45553dc5-54ea-4de9-914d-55b8fee477f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d112b1ea41c3d30a7cfa46d17db6aed0999eb0baaaf8a2f21c741d805a21451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d64j4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f76a1f655f871deb7e1364fcc5dbf6b564c4c386300bb49cb75f83518b64286e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d64j4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hbm26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:26Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.813538 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"402a6dc5-6bc7-4d7c-a909-9bb4ca48365d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d5b032497b592363a03f89cb31ba1b551f479004072dddffce540c41190ab81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a3fd02d55a10647428a05a9f53a1429a9827059c523d5deab3c4581c8f1b933\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11600ada13fbc7db1831de385202bc1f041efb656cc58aa49b0f71fe902db9bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fd39f5a05dec7a756dfd519f5df3ef7c77bbb5b8633c5af18a09ff8029cb079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:26Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.827767 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab21f7f0-7a5d-4eed-bd82-6c9d2e697eda\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ae3c8e412002f52fe150522d9bacb6415308ea41825b90d9d89010a7eca8764\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73fee4346d0af07556e8a1ea336818acea1f39aa8f703056b5954c7fac72aee1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdff2efc8ee6095385be74fbe1bf30926e526639dcc00b39fca8d5c9a0145d3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://742ee25d4d54068f173176593cafc678ec8502b53656fe9c969177f803836e29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://742ee25d4d54068f173176593cafc678ec8502b53656fe9c969177f803836e29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:26Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.841758 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:26Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.858266 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b38b8410bea95465ee58803c4160039bf7c082ca47893680da7469418f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27003b336112760daa2079f4819641850c27fbc8415b63b5094c86195291aca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:26Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.877936 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jh6vk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76a05d6-d9c7-4183-ad90-b3131f225841\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f57aea989227fdbd7bb59c51fade50d1076929f6288d060808b51bd1bdd7df04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f2659e2a914d3191daddbfb5ad109e487f03dabb559d324244e127add7d41d9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T15:16:24Z\\\",\\\"message\\\":\\\"2026-02-02T15:15:39+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_bbf1a3b4-24d2-4558-bc15-4414386de7b4\\\\n2026-02-02T15:15:39+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_bbf1a3b4-24d2-4558-bc15-4414386de7b4 to /host/opt/cni/bin/\\\\n2026-02-02T15:15:39Z [verbose] multus-daemon started\\\\n2026-02-02T15:15:39Z [verbose] Readiness Indicator file check\\\\n2026-02-02T15:16:24Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jh6vk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:26Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.894429 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83d0e810-ea5c-47aa-947b-7865b045bf94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6144eb32f88bc106f7724e48133b568a137c4f3fbbfe30c7fcef405d3ab735a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51318ad71b42f4cc8d376aa50e44580e68da89d83d80defb996329a2ee5c2ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zgsv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:26Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.902706 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.902774 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.902794 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.902823 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.902845 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:26Z","lastTransitionTime":"2026-02-02T15:16:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.907799 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5xgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56681968-f264-440e-aa12-d4f81879efe9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e250891b1065a38a448aa6abd300b37543e819e87f006b03cc37983cb643982a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdjjl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5xgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:26Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.920071 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7r9fd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4dfd44b-b363-4162-9f43-9a8dc4a3b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7r9fd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:26Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.938373 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7082da2-f2e7-4b73-ba35-504e5cd4c920\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd383d69b9b27d52fa8672ff6e0294a74876d676973bdc0ca95125fef5f2d085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037e379a109ce230098e3ca223fd1dea3149de3ebe601f56e2bb295f3093a3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6fdd19dac506f2d5d7eedc44ec2cda92a9381b73e91a45d949e4fcde4dfc283\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13965f5abcfb0a36c33310518955f192673e630a1aaba94b49400026de5bc2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://992f4198192cbe4b1c89d575aa4a20c29c1115fd5eaa8986c6ed70c1d6b143c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0202 15:15:31.010033 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 15:15:31.011243 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3543528698/tls.crt::/tmp/serving-cert-3543528698/tls.key\\\\\\\"\\\\nI0202 15:15:36.416658 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 15:15:36.428087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 15:15:36.428134 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 15:15:36.428243 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 15:15:36.428280 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 15:15:36.439678 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 15:15:36.439710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439717 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439722 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 15:15:36.439726 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 15:15:36.439729 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 15:15:36.439732 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 15:15:36.439753 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 15:15:36.450716 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://226f5949820d4f498dbf6fa251c6edf1c6356900bd946e6afae2c29042b1c835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:26Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.960548 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da08f9aecc070b31df2ddc1be7c6eb2ad8ef68a931e57ca2e5c5302f4e75751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:26Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.980341 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:26Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:26 crc kubenswrapper[4733]: I0202 15:16:26.990842 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca57dabf1d1ee5c6fc29291b1be687ca7e97c72650084b02067057fccac41a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:26Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.005320 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.005366 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.005379 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.005395 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.005407 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:27Z","lastTransitionTime":"2026-02-02T15:16:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.006340 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:27Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.107801 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.107850 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.107858 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.107871 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.107879 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:27Z","lastTransitionTime":"2026-02-02T15:16:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.211077 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.211121 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.211178 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.211197 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.211210 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:27Z","lastTransitionTime":"2026-02-02T15:16:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.238196 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-18 09:57:32.75761222 +0000 UTC Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.254745 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.254760 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.254809 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.254745 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:16:27 crc kubenswrapper[4733]: E0202 15:16:27.254890 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:16:27 crc kubenswrapper[4733]: E0202 15:16:27.254837 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:16:27 crc kubenswrapper[4733]: E0202 15:16:27.255035 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:16:27 crc kubenswrapper[4733]: E0202 15:16:27.255117 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.267852 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7082da2-f2e7-4b73-ba35-504e5cd4c920\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd383d69b9b27d52fa8672ff6e0294a74876d676973bdc0ca95125fef5f2d085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037e379a109ce230098e3ca223fd1dea3149de3ebe601f56e2bb295f3093a3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6fdd19dac506f2d5d7eedc44ec2cda92a9381b73e91a45d949e4fcde4dfc283\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13965f5abcfb0a36c33310518955f192673e630a1aaba94b49400026de5bc2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://992f4198192cbe4b1c89d575aa4a20c29c1115fd5eaa8986c6ed70c1d6b143c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0202 15:15:31.010033 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 15:15:31.011243 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3543528698/tls.crt::/tmp/serving-cert-3543528698/tls.key\\\\\\\"\\\\nI0202 15:15:36.416658 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 15:15:36.428087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 15:15:36.428134 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 15:15:36.428243 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 15:15:36.428280 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 15:15:36.439678 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 15:15:36.439710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439717 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439722 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 15:15:36.439726 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 15:15:36.439729 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 15:15:36.439732 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 15:15:36.439753 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 15:15:36.450716 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://226f5949820d4f498dbf6fa251c6edf1c6356900bd946e6afae2c29042b1c835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:27Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.284864 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da08f9aecc070b31df2ddc1be7c6eb2ad8ef68a931e57ca2e5c5302f4e75751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:27Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.296735 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:27Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.308115 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca57dabf1d1ee5c6fc29291b1be687ca7e97c72650084b02067057fccac41a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:27Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.317602 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.317670 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.317692 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.317719 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.317741 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:27Z","lastTransitionTime":"2026-02-02T15:16:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.320844 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:27Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.332035 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5xgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56681968-f264-440e-aa12-d4f81879efe9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e250891b1065a38a448aa6abd300b37543e819e87f006b03cc37983cb643982a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdjjl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5xgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:27Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.344679 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7r9fd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4dfd44b-b363-4162-9f43-9a8dc4a3b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7r9fd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:27Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.367533 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8206eca4-7d8b-4b59-a4c1-24631dd8cd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41d4a7bdc21f00b5039385d7227d8cb0f4ea3051d3b5065c71a49cb0bfad26a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c11b1859bf3bc0ec4d5e6a58d9becc1ac6333a1ed9820d83d634f1e93221c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c08b087d85be9bb83007c7e38a5de9182a5091716fb8a2ca5ded1b8191bebc40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fd4969f0f16c9aa44208d2bf578f77d9d50914a8cb1764507175aa2eb58e733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59175c8e5a7f3536168a69fd944bf3d536bcd7281eee96002343b7836a54401b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:27Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.384446 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nr6ps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e43cd902-791a-4042-bc08-cf03e075be32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2a89a5d1fc43696f33d313aee5dca316a249fdef6a23ff655e31f14cc449cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07e56782051c93b2a320aa539a42764b97647237d228bed2b6af48cee987c60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07e56782051c93b2a320aa539a42764b97647237d228bed2b6af48cee987c60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce850ac7bd4c2444b65d87232114d55b78d11c6f973537c615900092a45d55e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce850ac7bd4c2444b65d87232114d55b78d11c6f973537c615900092a45d55e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nr6ps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:27Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.410838 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5a81ada-e393-499a-aac6-7b40349958bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4acf2398e2e19bd54f482cb2273c54738e61627187860bb0b6daadddc2565b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831d85055203f153381eb51d98af0cd2e4c4a0309f837545ab3d855a680133e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1380122d1694c1286f1102e6396c6504b91f63c6eb157689dfc0ea7188c17f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5835cb53385359e9c58eaa5d4da2569c313e0fc265ad79a6c124e8cecb1a734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd6dc4bf5b9fd31bfbee6d9340fe981c0c5fca03a79cef9316340ff5a8b3d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db1db75e392ebea8b387b85cabc4276d553bf2c06694cf7f294a383909f32826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa356e9d1cf1236a6d325e762401b42848ed4c48bc4c538a5844e09de54d16c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa356e9d1cf1236a6d325e762401b42848ed4c48bc4c538a5844e09de54d16c1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T15:16:02Z\\\",\\\"message\\\":\\\"\\\\nI0202 15:16:02.301761 6382 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-5xgpn in node crc\\\\nI0202 15:16:02.301771 6382 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-8c597\\\\nI0202 15:16:02.301779 6382 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-5xgpn after 0 failed attempt(s)\\\\nI0202 15:16:02.301784 6382 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-8c597 in node crc\\\\nI0202 15:16:02.301793 6382 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-5xgpn\\\\nI0202 15:16:02.301794 6382 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-8c597 after 0 failed attempt(s)\\\\nI0202 15:16:02.301805 6382 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-8c597\\\\nI0202 15:16:02.301355 6382 services_controller.go:360] Finished syncing service metrics on namespace openshift-ingress-operator for network=default : 1.801637ms\\\\nI0202 15:16:02.301684 6382 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI0202 15:16:02.301833 6382 services_controller.go:356] Processing sync for service openshift-console-operator/metrics for network=default\\\\nI0202 15:16:02\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:16:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qxvkf_openshift-ovn-kubernetes(f5a81ada-e393-499a-aac6-7b40349958bb)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efaf2216e4c1750139f080a93b4b608737dab2883de5fddfd9a99248d517b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxvkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:27Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.420864 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.420894 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.420906 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.420921 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.420934 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:27Z","lastTransitionTime":"2026-02-02T15:16:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.421829 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8c597" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76cc68c3-9e90-45de-87b2-6ea1b68eadce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ac57831c1c880e64aaa7bedcf999133ee09fc5ecc5124b9d54ecedcc9a13c28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwr7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8c597\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:27Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.437334 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hbm26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45553dc5-54ea-4de9-914d-55b8fee477f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d112b1ea41c3d30a7cfa46d17db6aed0999eb0baaaf8a2f21c741d805a21451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d64j4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f76a1f655f871deb7e1364fcc5dbf6b564c4c386300bb49cb75f83518b64286e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d64j4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hbm26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:27Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.455219 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"402a6dc5-6bc7-4d7c-a909-9bb4ca48365d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d5b032497b592363a03f89cb31ba1b551f479004072dddffce540c41190ab81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a3fd02d55a10647428a05a9f53a1429a9827059c523d5deab3c4581c8f1b933\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11600ada13fbc7db1831de385202bc1f041efb656cc58aa49b0f71fe902db9bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fd39f5a05dec7a756dfd519f5df3ef7c77bbb5b8633c5af18a09ff8029cb079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:27Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.471148 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab21f7f0-7a5d-4eed-bd82-6c9d2e697eda\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ae3c8e412002f52fe150522d9bacb6415308ea41825b90d9d89010a7eca8764\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73fee4346d0af07556e8a1ea336818acea1f39aa8f703056b5954c7fac72aee1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdff2efc8ee6095385be74fbe1bf30926e526639dcc00b39fca8d5c9a0145d3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://742ee25d4d54068f173176593cafc678ec8502b53656fe9c969177f803836e29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://742ee25d4d54068f173176593cafc678ec8502b53656fe9c969177f803836e29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:27Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.487933 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:27Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.504613 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b38b8410bea95465ee58803c4160039bf7c082ca47893680da7469418f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27003b336112760daa2079f4819641850c27fbc8415b63b5094c86195291aca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:27Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.521031 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jh6vk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76a05d6-d9c7-4183-ad90-b3131f225841\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f57aea989227fdbd7bb59c51fade50d1076929f6288d060808b51bd1bdd7df04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f2659e2a914d3191daddbfb5ad109e487f03dabb559d324244e127add7d41d9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T15:16:24Z\\\",\\\"message\\\":\\\"2026-02-02T15:15:39+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_bbf1a3b4-24d2-4558-bc15-4414386de7b4\\\\n2026-02-02T15:15:39+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_bbf1a3b4-24d2-4558-bc15-4414386de7b4 to /host/opt/cni/bin/\\\\n2026-02-02T15:15:39Z [verbose] multus-daemon started\\\\n2026-02-02T15:15:39Z [verbose] Readiness Indicator file check\\\\n2026-02-02T15:16:24Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jh6vk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:27Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.522578 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.522602 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.522613 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.522628 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.522640 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:27Z","lastTransitionTime":"2026-02-02T15:16:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.534879 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83d0e810-ea5c-47aa-947b-7865b045bf94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6144eb32f88bc106f7724e48133b568a137c4f3fbbfe30c7fcef405d3ab735a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51318ad71b42f4cc8d376aa50e44580e68da89d83d80defb996329a2ee5c2ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zgsv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:27Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.625201 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.625333 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.625391 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.625464 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.625526 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:27Z","lastTransitionTime":"2026-02-02T15:16:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.727739 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.727797 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.727814 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.727836 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.727880 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:27Z","lastTransitionTime":"2026-02-02T15:16:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.738786 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.738828 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.738843 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.738860 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.738874 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:27Z","lastTransitionTime":"2026-02-02T15:16:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:27 crc kubenswrapper[4733]: E0202 15:16:27.751960 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd742269-802e-45ec-a88e-17aee17c574e\\\",\\\"systemUUID\\\":\\\"91b64b75-3cd9-4fb9-93f0-dbd3bcdcaec4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:27Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.756089 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.756124 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.756132 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.756145 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.756154 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:27Z","lastTransitionTime":"2026-02-02T15:16:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:27 crc kubenswrapper[4733]: E0202 15:16:27.771127 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd742269-802e-45ec-a88e-17aee17c574e\\\",\\\"systemUUID\\\":\\\"91b64b75-3cd9-4fb9-93f0-dbd3bcdcaec4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:27Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.774518 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.774591 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.774610 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.774635 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.774654 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:27Z","lastTransitionTime":"2026-02-02T15:16:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:27 crc kubenswrapper[4733]: E0202 15:16:27.786021 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd742269-802e-45ec-a88e-17aee17c574e\\\",\\\"systemUUID\\\":\\\"91b64b75-3cd9-4fb9-93f0-dbd3bcdcaec4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:27Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.788926 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.788947 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.788955 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.788968 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.788976 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:27Z","lastTransitionTime":"2026-02-02T15:16:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:27 crc kubenswrapper[4733]: E0202 15:16:27.798331 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd742269-802e-45ec-a88e-17aee17c574e\\\",\\\"systemUUID\\\":\\\"91b64b75-3cd9-4fb9-93f0-dbd3bcdcaec4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:27Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.801199 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.801227 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.801262 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.801275 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.801284 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:27Z","lastTransitionTime":"2026-02-02T15:16:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:27 crc kubenswrapper[4733]: E0202 15:16:27.812530 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd742269-802e-45ec-a88e-17aee17c574e\\\",\\\"systemUUID\\\":\\\"91b64b75-3cd9-4fb9-93f0-dbd3bcdcaec4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:27Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:27 crc kubenswrapper[4733]: E0202 15:16:27.812630 4733 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.830146 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.830232 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.830249 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.830273 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.830290 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:27Z","lastTransitionTime":"2026-02-02T15:16:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.933444 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.933490 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.933507 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.933529 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:27 crc kubenswrapper[4733]: I0202 15:16:27.933545 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:27Z","lastTransitionTime":"2026-02-02T15:16:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:28 crc kubenswrapper[4733]: I0202 15:16:28.036186 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:28 crc kubenswrapper[4733]: I0202 15:16:28.036222 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:28 crc kubenswrapper[4733]: I0202 15:16:28.036240 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:28 crc kubenswrapper[4733]: I0202 15:16:28.036261 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:28 crc kubenswrapper[4733]: I0202 15:16:28.036279 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:28Z","lastTransitionTime":"2026-02-02T15:16:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:28 crc kubenswrapper[4733]: I0202 15:16:28.138403 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:28 crc kubenswrapper[4733]: I0202 15:16:28.138459 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:28 crc kubenswrapper[4733]: I0202 15:16:28.138479 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:28 crc kubenswrapper[4733]: I0202 15:16:28.138503 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:28 crc kubenswrapper[4733]: I0202 15:16:28.138521 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:28Z","lastTransitionTime":"2026-02-02T15:16:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:28 crc kubenswrapper[4733]: I0202 15:16:28.239284 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 09:12:33.142503725 +0000 UTC Feb 02 15:16:28 crc kubenswrapper[4733]: I0202 15:16:28.241269 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:28 crc kubenswrapper[4733]: I0202 15:16:28.241319 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:28 crc kubenswrapper[4733]: I0202 15:16:28.241334 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:28 crc kubenswrapper[4733]: I0202 15:16:28.241352 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:28 crc kubenswrapper[4733]: I0202 15:16:28.241364 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:28Z","lastTransitionTime":"2026-02-02T15:16:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:28 crc kubenswrapper[4733]: I0202 15:16:28.344401 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:28 crc kubenswrapper[4733]: I0202 15:16:28.344434 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:28 crc kubenswrapper[4733]: I0202 15:16:28.344445 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:28 crc kubenswrapper[4733]: I0202 15:16:28.344459 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:28 crc kubenswrapper[4733]: I0202 15:16:28.344470 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:28Z","lastTransitionTime":"2026-02-02T15:16:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:28 crc kubenswrapper[4733]: I0202 15:16:28.446797 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:28 crc kubenswrapper[4733]: I0202 15:16:28.446877 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:28 crc kubenswrapper[4733]: I0202 15:16:28.446900 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:28 crc kubenswrapper[4733]: I0202 15:16:28.446934 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:28 crc kubenswrapper[4733]: I0202 15:16:28.446956 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:28Z","lastTransitionTime":"2026-02-02T15:16:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:28 crc kubenswrapper[4733]: I0202 15:16:28.550607 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:28 crc kubenswrapper[4733]: I0202 15:16:28.550678 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:28 crc kubenswrapper[4733]: I0202 15:16:28.550699 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:28 crc kubenswrapper[4733]: I0202 15:16:28.550724 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:28 crc kubenswrapper[4733]: I0202 15:16:28.550744 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:28Z","lastTransitionTime":"2026-02-02T15:16:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:28 crc kubenswrapper[4733]: I0202 15:16:28.653721 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:28 crc kubenswrapper[4733]: I0202 15:16:28.653786 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:28 crc kubenswrapper[4733]: I0202 15:16:28.653797 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:28 crc kubenswrapper[4733]: I0202 15:16:28.653815 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:28 crc kubenswrapper[4733]: I0202 15:16:28.653829 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:28Z","lastTransitionTime":"2026-02-02T15:16:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:28 crc kubenswrapper[4733]: I0202 15:16:28.757128 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:28 crc kubenswrapper[4733]: I0202 15:16:28.757201 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:28 crc kubenswrapper[4733]: I0202 15:16:28.757218 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:28 crc kubenswrapper[4733]: I0202 15:16:28.757240 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:28 crc kubenswrapper[4733]: I0202 15:16:28.757257 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:28Z","lastTransitionTime":"2026-02-02T15:16:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:28 crc kubenswrapper[4733]: I0202 15:16:28.860282 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:28 crc kubenswrapper[4733]: I0202 15:16:28.860343 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:28 crc kubenswrapper[4733]: I0202 15:16:28.860364 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:28 crc kubenswrapper[4733]: I0202 15:16:28.860392 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:28 crc kubenswrapper[4733]: I0202 15:16:28.860413 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:28Z","lastTransitionTime":"2026-02-02T15:16:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:28 crc kubenswrapper[4733]: I0202 15:16:28.963009 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:28 crc kubenswrapper[4733]: I0202 15:16:28.963045 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:28 crc kubenswrapper[4733]: I0202 15:16:28.963057 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:28 crc kubenswrapper[4733]: I0202 15:16:28.963078 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:28 crc kubenswrapper[4733]: I0202 15:16:28.963089 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:28Z","lastTransitionTime":"2026-02-02T15:16:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.069961 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.070032 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.070050 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.070075 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.070099 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:29Z","lastTransitionTime":"2026-02-02T15:16:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.172701 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.172756 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.172774 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.172798 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.172814 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:29Z","lastTransitionTime":"2026-02-02T15:16:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.240332 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-11 05:07:20.139045656 +0000 UTC Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.254677 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.254764 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.255297 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:16:29 crc kubenswrapper[4733]: E0202 15:16:29.255398 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.255536 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:16:29 crc kubenswrapper[4733]: E0202 15:16:29.255646 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:16:29 crc kubenswrapper[4733]: E0202 15:16:29.255717 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.255792 4733 scope.go:117] "RemoveContainer" containerID="fa356e9d1cf1236a6d325e762401b42848ed4c48bc4c538a5844e09de54d16c1" Feb 02 15:16:29 crc kubenswrapper[4733]: E0202 15:16:29.255835 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.275074 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.275451 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.275468 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.275489 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.275505 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:29Z","lastTransitionTime":"2026-02-02T15:16:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.380848 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.380891 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.380906 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.380929 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.380947 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:29Z","lastTransitionTime":"2026-02-02T15:16:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.484575 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.484621 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.484639 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.484662 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.484680 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:29Z","lastTransitionTime":"2026-02-02T15:16:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.586565 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.586601 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.586613 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.586629 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.586641 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:29Z","lastTransitionTime":"2026-02-02T15:16:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.688932 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.688996 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.689009 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.689023 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.689332 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:29Z","lastTransitionTime":"2026-02-02T15:16:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.728840 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxvkf_f5a81ada-e393-499a-aac6-7b40349958bb/ovnkube-controller/2.log" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.734120 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" event={"ID":"f5a81ada-e393-499a-aac6-7b40349958bb","Type":"ContainerStarted","Data":"55b2c8f5f7c8b659a1b54bb592229f81596804f972fc4658e1f867054721257d"} Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.734959 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.751568 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8206eca4-7d8b-4b59-a4c1-24631dd8cd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41d4a7bdc21f00b5039385d7227d8cb0f4ea3051d3b5065c71a49cb0bfad26a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c11b1859bf3bc0ec4d5e6a58d9becc1ac6333a1ed9820d83d634f1e93221c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c08b087d85be9bb83007c7e38a5de9182a5091716fb8a2ca5ded1b8191bebc40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fd4969f0f16c9aa44208d2bf578f77d9d50914a8cb1764507175aa2eb58e733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59175c8e5a7f3536168a69fd944bf3d536bcd7281eee96002343b7836a54401b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:29Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.764679 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nr6ps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e43cd902-791a-4042-bc08-cf03e075be32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2a89a5d1fc43696f33d313aee5dca316a249fdef6a23ff655e31f14cc449cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07e56782051c93b2a320aa539a42764b97647237d228bed2b6af48cee987c60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07e56782051c93b2a320aa539a42764b97647237d228bed2b6af48cee987c60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce850ac7bd4c2444b65d87232114d55b78d11c6f973537c615900092a45d55e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce850ac7bd4c2444b65d87232114d55b78d11c6f973537c615900092a45d55e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nr6ps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:29Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.781414 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5a81ada-e393-499a-aac6-7b40349958bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4acf2398e2e19bd54f482cb2273c54738e61627187860bb0b6daadddc2565b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831d85055203f153381eb51d98af0cd2e4c4a0309f837545ab3d855a680133e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1380122d1694c1286f1102e6396c6504b91f63c6eb157689dfc0ea7188c17f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5835cb53385359e9c58eaa5d4da2569c313e0fc265ad79a6c124e8cecb1a734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd6dc4bf5b9fd31bfbee6d9340fe981c0c5fca03a79cef9316340ff5a8b3d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db1db75e392ebea8b387b85cabc4276d553bf2c06694cf7f294a383909f32826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55b2c8f5f7c8b659a1b54bb592229f81596804f972fc4658e1f867054721257d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa356e9d1cf1236a6d325e762401b42848ed4c48bc4c538a5844e09de54d16c1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T15:16:02Z\\\",\\\"message\\\":\\\"\\\\nI0202 15:16:02.301761 6382 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-5xgpn in node crc\\\\nI0202 15:16:02.301771 6382 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-8c597\\\\nI0202 15:16:02.301779 6382 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-5xgpn after 0 failed attempt(s)\\\\nI0202 15:16:02.301784 6382 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-8c597 in node crc\\\\nI0202 15:16:02.301793 6382 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-5xgpn\\\\nI0202 15:16:02.301794 6382 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-8c597 after 0 failed attempt(s)\\\\nI0202 15:16:02.301805 6382 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-8c597\\\\nI0202 15:16:02.301355 6382 services_controller.go:360] Finished syncing service metrics on namespace openshift-ingress-operator for network=default : 1.801637ms\\\\nI0202 15:16:02.301684 6382 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI0202 15:16:02.301833 6382 services_controller.go:356] Processing sync for service openshift-console-operator/metrics for network=default\\\\nI0202 15:16:02\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:16:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efaf2216e4c1750139f080a93b4b608737dab2883de5fddfd9a99248d517b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxvkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:29Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.790902 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8c597" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76cc68c3-9e90-45de-87b2-6ea1b68eadce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ac57831c1c880e64aaa7bedcf999133ee09fc5ecc5124b9d54ecedcc9a13c28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwr7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8c597\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:29Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.792094 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.792135 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.792149 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.792185 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.792199 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:29Z","lastTransitionTime":"2026-02-02T15:16:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.800619 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hbm26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45553dc5-54ea-4de9-914d-55b8fee477f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d112b1ea41c3d30a7cfa46d17db6aed0999eb0baaaf8a2f21c741d805a21451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d64j4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f76a1f655f871deb7e1364fcc5dbf6b564c4c386300bb49cb75f83518b64286e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d64j4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hbm26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:29Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.812127 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"402a6dc5-6bc7-4d7c-a909-9bb4ca48365d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d5b032497b592363a03f89cb31ba1b551f479004072dddffce540c41190ab81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a3fd02d55a10647428a05a9f53a1429a9827059c523d5deab3c4581c8f1b933\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11600ada13fbc7db1831de385202bc1f041efb656cc58aa49b0f71fe902db9bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fd39f5a05dec7a756dfd519f5df3ef7c77bbb5b8633c5af18a09ff8029cb079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:29Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.825818 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab21f7f0-7a5d-4eed-bd82-6c9d2e697eda\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ae3c8e412002f52fe150522d9bacb6415308ea41825b90d9d89010a7eca8764\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73fee4346d0af07556e8a1ea336818acea1f39aa8f703056b5954c7fac72aee1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdff2efc8ee6095385be74fbe1bf30926e526639dcc00b39fca8d5c9a0145d3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://742ee25d4d54068f173176593cafc678ec8502b53656fe9c969177f803836e29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://742ee25d4d54068f173176593cafc678ec8502b53656fe9c969177f803836e29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:29Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.838678 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:29Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.849643 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b38b8410bea95465ee58803c4160039bf7c082ca47893680da7469418f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27003b336112760daa2079f4819641850c27fbc8415b63b5094c86195291aca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:29Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.859640 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jh6vk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76a05d6-d9c7-4183-ad90-b3131f225841\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f57aea989227fdbd7bb59c51fade50d1076929f6288d060808b51bd1bdd7df04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f2659e2a914d3191daddbfb5ad109e487f03dabb559d324244e127add7d41d9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T15:16:24Z\\\",\\\"message\\\":\\\"2026-02-02T15:15:39+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_bbf1a3b4-24d2-4558-bc15-4414386de7b4\\\\n2026-02-02T15:15:39+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_bbf1a3b4-24d2-4558-bc15-4414386de7b4 to /host/opt/cni/bin/\\\\n2026-02-02T15:15:39Z [verbose] multus-daemon started\\\\n2026-02-02T15:15:39Z [verbose] Readiness Indicator file check\\\\n2026-02-02T15:16:24Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jh6vk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:29Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.869060 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83d0e810-ea5c-47aa-947b-7865b045bf94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6144eb32f88bc106f7724e48133b568a137c4f3fbbfe30c7fcef405d3ab735a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51318ad71b42f4cc8d376aa50e44580e68da89d83d80defb996329a2ee5c2ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zgsv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:29Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.882283 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7082da2-f2e7-4b73-ba35-504e5cd4c920\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd383d69b9b27d52fa8672ff6e0294a74876d676973bdc0ca95125fef5f2d085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037e379a109ce230098e3ca223fd1dea3149de3ebe601f56e2bb295f3093a3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6fdd19dac506f2d5d7eedc44ec2cda92a9381b73e91a45d949e4fcde4dfc283\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13965f5abcfb0a36c33310518955f192673e630a1aaba94b49400026de5bc2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://992f4198192cbe4b1c89d575aa4a20c29c1115fd5eaa8986c6ed70c1d6b143c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0202 15:15:31.010033 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 15:15:31.011243 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3543528698/tls.crt::/tmp/serving-cert-3543528698/tls.key\\\\\\\"\\\\nI0202 15:15:36.416658 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 15:15:36.428087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 15:15:36.428134 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 15:15:36.428243 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 15:15:36.428280 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 15:15:36.439678 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 15:15:36.439710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439717 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439722 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 15:15:36.439726 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 15:15:36.439729 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 15:15:36.439732 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 15:15:36.439753 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 15:15:36.450716 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://226f5949820d4f498dbf6fa251c6edf1c6356900bd946e6afae2c29042b1c835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:29Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.893214 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da08f9aecc070b31df2ddc1be7c6eb2ad8ef68a931e57ca2e5c5302f4e75751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:29Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.894352 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.894392 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.894404 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.894422 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.894432 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:29Z","lastTransitionTime":"2026-02-02T15:16:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.903548 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:29Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.914971 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca57dabf1d1ee5c6fc29291b1be687ca7e97c72650084b02067057fccac41a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:29Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.925781 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:29Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.934266 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5xgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56681968-f264-440e-aa12-d4f81879efe9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e250891b1065a38a448aa6abd300b37543e819e87f006b03cc37983cb643982a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdjjl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5xgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:29Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.942864 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7r9fd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4dfd44b-b363-4162-9f43-9a8dc4a3b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7r9fd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:29Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.996755 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.996785 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.996793 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.996806 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:29 crc kubenswrapper[4733]: I0202 15:16:29.996815 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:29Z","lastTransitionTime":"2026-02-02T15:16:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.099244 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.099284 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.099294 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.099307 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.099315 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:30Z","lastTransitionTime":"2026-02-02T15:16:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.201065 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.201121 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.201133 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.201202 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.201221 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:30Z","lastTransitionTime":"2026-02-02T15:16:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.241473 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-30 18:36:28.715676828 +0000 UTC Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.303974 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.304032 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.304049 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.304074 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.304091 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:30Z","lastTransitionTime":"2026-02-02T15:16:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.406507 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.406569 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.406588 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.406613 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.406629 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:30Z","lastTransitionTime":"2026-02-02T15:16:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.509278 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.509343 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.509354 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.509370 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.509382 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:30Z","lastTransitionTime":"2026-02-02T15:16:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.612353 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.612410 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.612425 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.612447 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.612461 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:30Z","lastTransitionTime":"2026-02-02T15:16:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.714525 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.714552 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.714563 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.714576 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.714586 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:30Z","lastTransitionTime":"2026-02-02T15:16:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.739544 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxvkf_f5a81ada-e393-499a-aac6-7b40349958bb/ovnkube-controller/3.log" Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.740454 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxvkf_f5a81ada-e393-499a-aac6-7b40349958bb/ovnkube-controller/2.log" Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.744041 4733 generic.go:334] "Generic (PLEG): container finished" podID="f5a81ada-e393-499a-aac6-7b40349958bb" containerID="55b2c8f5f7c8b659a1b54bb592229f81596804f972fc4658e1f867054721257d" exitCode=1 Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.744098 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" event={"ID":"f5a81ada-e393-499a-aac6-7b40349958bb","Type":"ContainerDied","Data":"55b2c8f5f7c8b659a1b54bb592229f81596804f972fc4658e1f867054721257d"} Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.744146 4733 scope.go:117] "RemoveContainer" containerID="fa356e9d1cf1236a6d325e762401b42848ed4c48bc4c538a5844e09de54d16c1" Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.744851 4733 scope.go:117] "RemoveContainer" containerID="55b2c8f5f7c8b659a1b54bb592229f81596804f972fc4658e1f867054721257d" Feb 02 15:16:30 crc kubenswrapper[4733]: E0202 15:16:30.745035 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-qxvkf_openshift-ovn-kubernetes(f5a81ada-e393-499a-aac6-7b40349958bb)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.761704 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jh6vk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76a05d6-d9c7-4183-ad90-b3131f225841\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f57aea989227fdbd7bb59c51fade50d1076929f6288d060808b51bd1bdd7df04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f2659e2a914d3191daddbfb5ad109e487f03dabb559d324244e127add7d41d9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T15:16:24Z\\\",\\\"message\\\":\\\"2026-02-02T15:15:39+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_bbf1a3b4-24d2-4558-bc15-4414386de7b4\\\\n2026-02-02T15:15:39+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_bbf1a3b4-24d2-4558-bc15-4414386de7b4 to /host/opt/cni/bin/\\\\n2026-02-02T15:15:39Z [verbose] multus-daemon started\\\\n2026-02-02T15:15:39Z [verbose] Readiness Indicator file check\\\\n2026-02-02T15:16:24Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jh6vk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:30Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.776226 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83d0e810-ea5c-47aa-947b-7865b045bf94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6144eb32f88bc106f7724e48133b568a137c4f3fbbfe30c7fcef405d3ab735a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51318ad71b42f4cc8d376aa50e44580e68da89d83d80defb996329a2ee5c2ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zgsv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:30Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.789043 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:30Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.801629 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b38b8410bea95465ee58803c4160039bf7c082ca47893680da7469418f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27003b336112760daa2079f4819641850c27fbc8415b63b5094c86195291aca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:30Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.813358 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da08f9aecc070b31df2ddc1be7c6eb2ad8ef68a931e57ca2e5c5302f4e75751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:30Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.816572 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.816604 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.816616 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.816630 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.816640 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:30Z","lastTransitionTime":"2026-02-02T15:16:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.825307 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:30Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.834774 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca57dabf1d1ee5c6fc29291b1be687ca7e97c72650084b02067057fccac41a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:30Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.848877 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:30Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.883086 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5xgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56681968-f264-440e-aa12-d4f81879efe9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e250891b1065a38a448aa6abd300b37543e819e87f006b03cc37983cb643982a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdjjl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5xgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:30Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.904081 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7r9fd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4dfd44b-b363-4162-9f43-9a8dc4a3b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7r9fd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:30Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.918859 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.918894 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.918903 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.918917 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.918925 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:30Z","lastTransitionTime":"2026-02-02T15:16:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.928728 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7082da2-f2e7-4b73-ba35-504e5cd4c920\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd383d69b9b27d52fa8672ff6e0294a74876d676973bdc0ca95125fef5f2d085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037e379a109ce230098e3ca223fd1dea3149de3ebe601f56e2bb295f3093a3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6fdd19dac506f2d5d7eedc44ec2cda92a9381b73e91a45d949e4fcde4dfc283\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13965f5abcfb0a36c33310518955f192673e630a1aaba94b49400026de5bc2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://992f4198192cbe4b1c89d575aa4a20c29c1115fd5eaa8986c6ed70c1d6b143c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0202 15:15:31.010033 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 15:15:31.011243 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3543528698/tls.crt::/tmp/serving-cert-3543528698/tls.key\\\\\\\"\\\\nI0202 15:15:36.416658 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 15:15:36.428087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 15:15:36.428134 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 15:15:36.428243 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 15:15:36.428280 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 15:15:36.439678 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 15:15:36.439710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439717 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439722 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 15:15:36.439726 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 15:15:36.439729 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 15:15:36.439732 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 15:15:36.439753 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 15:15:36.450716 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://226f5949820d4f498dbf6fa251c6edf1c6356900bd946e6afae2c29042b1c835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:30Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.950140 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5a81ada-e393-499a-aac6-7b40349958bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4acf2398e2e19bd54f482cb2273c54738e61627187860bb0b6daadddc2565b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831d85055203f153381eb51d98af0cd2e4c4a0309f837545ab3d855a680133e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1380122d1694c1286f1102e6396c6504b91f63c6eb157689dfc0ea7188c17f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5835cb53385359e9c58eaa5d4da2569c313e0fc265ad79a6c124e8cecb1a734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd6dc4bf5b9fd31bfbee6d9340fe981c0c5fca03a79cef9316340ff5a8b3d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db1db75e392ebea8b387b85cabc4276d553bf2c06694cf7f294a383909f32826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55b2c8f5f7c8b659a1b54bb592229f81596804f972fc4658e1f867054721257d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa356e9d1cf1236a6d325e762401b42848ed4c48bc4c538a5844e09de54d16c1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T15:16:02Z\\\",\\\"message\\\":\\\"\\\\nI0202 15:16:02.301761 6382 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-5xgpn in node crc\\\\nI0202 15:16:02.301771 6382 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-8c597\\\\nI0202 15:16:02.301779 6382 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-5xgpn after 0 failed attempt(s)\\\\nI0202 15:16:02.301784 6382 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-8c597 in node crc\\\\nI0202 15:16:02.301793 6382 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-5xgpn\\\\nI0202 15:16:02.301794 6382 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-8c597 after 0 failed attempt(s)\\\\nI0202 15:16:02.301805 6382 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-8c597\\\\nI0202 15:16:02.301355 6382 services_controller.go:360] Finished syncing service metrics on namespace openshift-ingress-operator for network=default : 1.801637ms\\\\nI0202 15:16:02.301684 6382 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI0202 15:16:02.301833 6382 services_controller.go:356] Processing sync for service openshift-console-operator/metrics for network=default\\\\nI0202 15:16:02\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:16:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55b2c8f5f7c8b659a1b54bb592229f81596804f972fc4658e1f867054721257d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T15:16:30Z\\\",\\\"message\\\":\\\"lse hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.138:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {97419c58-41c7-41d7-a137-a446f0c7eeb3}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0202 15:16:30.146591 6766 obj_retry.go:551] Creating *factory.egressNode crc took: 3.786809ms\\\\nI0202 15:16:30.147325 6766 factory.go:1336] Added *v1.Node event handler 7\\\\nI0202 15:16:30.147474 6766 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI0202 15:16:30.147563 6766 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-operators\\\\\\\"}\\\\nI0202 15:16:30.147582 6766 services_controller.go:360] Finished syncing service redhat-operators on namespace openshift-marketplace for network=default : 6.641214ms\\\\nI0202 15:16:30.149227 6766 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0202 15:16:30.149857 6766 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0202 15:16:30.150220 6766 ovnkube.go:599] Stopped ovnkube\\\\nI0202 15:16:30.150339 6766 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0202 15:16:30.150729 6766 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:16:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efaf2216e4c1750139f080a93b4b608737dab2883de5fddfd9a99248d517b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxvkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:30Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.958870 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8c597" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76cc68c3-9e90-45de-87b2-6ea1b68eadce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ac57831c1c880e64aaa7bedcf999133ee09fc5ecc5124b9d54ecedcc9a13c28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwr7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8c597\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:30Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.969491 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hbm26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45553dc5-54ea-4de9-914d-55b8fee477f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d112b1ea41c3d30a7cfa46d17db6aed0999eb0baaaf8a2f21c741d805a21451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d64j4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f76a1f655f871deb7e1364fcc5dbf6b564c4c386300bb49cb75f83518b64286e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d64j4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hbm26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:30Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:30 crc kubenswrapper[4733]: I0202 15:16:30.989994 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8206eca4-7d8b-4b59-a4c1-24631dd8cd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41d4a7bdc21f00b5039385d7227d8cb0f4ea3051d3b5065c71a49cb0bfad26a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c11b1859bf3bc0ec4d5e6a58d9becc1ac6333a1ed9820d83d634f1e93221c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c08b087d85be9bb83007c7e38a5de9182a5091716fb8a2ca5ded1b8191bebc40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fd4969f0f16c9aa44208d2bf578f77d9d50914a8cb1764507175aa2eb58e733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59175c8e5a7f3536168a69fd944bf3d536bcd7281eee96002343b7836a54401b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:30Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.004355 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nr6ps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e43cd902-791a-4042-bc08-cf03e075be32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2a89a5d1fc43696f33d313aee5dca316a249fdef6a23ff655e31f14cc449cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07e56782051c93b2a320aa539a42764b97647237d228bed2b6af48cee987c60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07e56782051c93b2a320aa539a42764b97647237d228bed2b6af48cee987c60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce850ac7bd4c2444b65d87232114d55b78d11c6f973537c615900092a45d55e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce850ac7bd4c2444b65d87232114d55b78d11c6f973537c615900092a45d55e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nr6ps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:31Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.019505 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"402a6dc5-6bc7-4d7c-a909-9bb4ca48365d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d5b032497b592363a03f89cb31ba1b551f479004072dddffce540c41190ab81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a3fd02d55a10647428a05a9f53a1429a9827059c523d5deab3c4581c8f1b933\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11600ada13fbc7db1831de385202bc1f041efb656cc58aa49b0f71fe902db9bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fd39f5a05dec7a756dfd519f5df3ef7c77bbb5b8633c5af18a09ff8029cb079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:31Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.021783 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.021838 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.021855 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.021878 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.021896 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:31Z","lastTransitionTime":"2026-02-02T15:16:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.030812 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab21f7f0-7a5d-4eed-bd82-6c9d2e697eda\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ae3c8e412002f52fe150522d9bacb6415308ea41825b90d9d89010a7eca8764\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73fee4346d0af07556e8a1ea336818acea1f39aa8f703056b5954c7fac72aee1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdff2efc8ee6095385be74fbe1bf30926e526639dcc00b39fca8d5c9a0145d3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://742ee25d4d54068f173176593cafc678ec8502b53656fe9c969177f803836e29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://742ee25d4d54068f173176593cafc678ec8502b53656fe9c969177f803836e29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:31Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.124931 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.125005 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.125024 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.125048 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.125066 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:31Z","lastTransitionTime":"2026-02-02T15:16:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.228093 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.228146 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.228190 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.228216 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.228233 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:31Z","lastTransitionTime":"2026-02-02T15:16:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.242539 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 22:09:38.280760215 +0000 UTC Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.254081 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:16:31 crc kubenswrapper[4733]: E0202 15:16:31.254281 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.254260 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.254339 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.254346 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:16:31 crc kubenswrapper[4733]: E0202 15:16:31.254432 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:16:31 crc kubenswrapper[4733]: E0202 15:16:31.254542 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:16:31 crc kubenswrapper[4733]: E0202 15:16:31.254618 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.331445 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.331511 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.331531 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.331555 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.331572 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:31Z","lastTransitionTime":"2026-02-02T15:16:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.435746 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.435824 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.435848 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.435874 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.435891 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:31Z","lastTransitionTime":"2026-02-02T15:16:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.541258 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.541301 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.541310 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.541325 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.541335 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:31Z","lastTransitionTime":"2026-02-02T15:16:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.644978 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.645042 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.645063 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.645086 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.645103 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:31Z","lastTransitionTime":"2026-02-02T15:16:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.747573 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.747640 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.747660 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.747687 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.747708 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:31Z","lastTransitionTime":"2026-02-02T15:16:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.752038 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxvkf_f5a81ada-e393-499a-aac6-7b40349958bb/ovnkube-controller/3.log" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.758659 4733 scope.go:117] "RemoveContainer" containerID="55b2c8f5f7c8b659a1b54bb592229f81596804f972fc4658e1f867054721257d" Feb 02 15:16:31 crc kubenswrapper[4733]: E0202 15:16:31.758927 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-qxvkf_openshift-ovn-kubernetes(f5a81ada-e393-499a-aac6-7b40349958bb)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.779536 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:31Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.800726 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b38b8410bea95465ee58803c4160039bf7c082ca47893680da7469418f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27003b336112760daa2079f4819641850c27fbc8415b63b5094c86195291aca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:31Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.822096 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jh6vk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76a05d6-d9c7-4183-ad90-b3131f225841\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f57aea989227fdbd7bb59c51fade50d1076929f6288d060808b51bd1bdd7df04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f2659e2a914d3191daddbfb5ad109e487f03dabb559d324244e127add7d41d9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T15:16:24Z\\\",\\\"message\\\":\\\"2026-02-02T15:15:39+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_bbf1a3b4-24d2-4558-bc15-4414386de7b4\\\\n2026-02-02T15:15:39+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_bbf1a3b4-24d2-4558-bc15-4414386de7b4 to /host/opt/cni/bin/\\\\n2026-02-02T15:15:39Z [verbose] multus-daemon started\\\\n2026-02-02T15:15:39Z [verbose] Readiness Indicator file check\\\\n2026-02-02T15:16:24Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jh6vk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:31Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.841382 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83d0e810-ea5c-47aa-947b-7865b045bf94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6144eb32f88bc106f7724e48133b568a137c4f3fbbfe30c7fcef405d3ab735a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51318ad71b42f4cc8d376aa50e44580e68da89d83d80defb996329a2ee5c2ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zgsv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:31Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.858046 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.858120 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.858148 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.858221 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.858251 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:31Z","lastTransitionTime":"2026-02-02T15:16:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.862926 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:31Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.881307 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5xgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56681968-f264-440e-aa12-d4f81879efe9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e250891b1065a38a448aa6abd300b37543e819e87f006b03cc37983cb643982a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdjjl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5xgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:31Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.900212 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7r9fd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4dfd44b-b363-4162-9f43-9a8dc4a3b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7r9fd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:31Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.923816 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7082da2-f2e7-4b73-ba35-504e5cd4c920\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd383d69b9b27d52fa8672ff6e0294a74876d676973bdc0ca95125fef5f2d085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037e379a109ce230098e3ca223fd1dea3149de3ebe601f56e2bb295f3093a3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6fdd19dac506f2d5d7eedc44ec2cda92a9381b73e91a45d949e4fcde4dfc283\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13965f5abcfb0a36c33310518955f192673e630a1aaba94b49400026de5bc2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://992f4198192cbe4b1c89d575aa4a20c29c1115fd5eaa8986c6ed70c1d6b143c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0202 15:15:31.010033 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 15:15:31.011243 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3543528698/tls.crt::/tmp/serving-cert-3543528698/tls.key\\\\\\\"\\\\nI0202 15:15:36.416658 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 15:15:36.428087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 15:15:36.428134 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 15:15:36.428243 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 15:15:36.428280 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 15:15:36.439678 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 15:15:36.439710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439717 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439722 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 15:15:36.439726 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 15:15:36.439729 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 15:15:36.439732 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 15:15:36.439753 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 15:15:36.450716 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://226f5949820d4f498dbf6fa251c6edf1c6356900bd946e6afae2c29042b1c835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:31Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.945061 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da08f9aecc070b31df2ddc1be7c6eb2ad8ef68a931e57ca2e5c5302f4e75751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:31Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.961360 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.961403 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.961419 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.961442 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.961458 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:31Z","lastTransitionTime":"2026-02-02T15:16:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.965244 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:31Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:31 crc kubenswrapper[4733]: I0202 15:16:31.990895 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca57dabf1d1ee5c6fc29291b1be687ca7e97c72650084b02067057fccac41a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:31Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.027770 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8206eca4-7d8b-4b59-a4c1-24631dd8cd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41d4a7bdc21f00b5039385d7227d8cb0f4ea3051d3b5065c71a49cb0bfad26a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c11b1859bf3bc0ec4d5e6a58d9becc1ac6333a1ed9820d83d634f1e93221c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c08b087d85be9bb83007c7e38a5de9182a5091716fb8a2ca5ded1b8191bebc40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fd4969f0f16c9aa44208d2bf578f77d9d50914a8cb1764507175aa2eb58e733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59175c8e5a7f3536168a69fd944bf3d536bcd7281eee96002343b7836a54401b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:32Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.051224 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nr6ps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e43cd902-791a-4042-bc08-cf03e075be32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2a89a5d1fc43696f33d313aee5dca316a249fdef6a23ff655e31f14cc449cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07e56782051c93b2a320aa539a42764b97647237d228bed2b6af48cee987c60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07e56782051c93b2a320aa539a42764b97647237d228bed2b6af48cee987c60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce850ac7bd4c2444b65d87232114d55b78d11c6f973537c615900092a45d55e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce850ac7bd4c2444b65d87232114d55b78d11c6f973537c615900092a45d55e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nr6ps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:32Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.065294 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.065360 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.065377 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.065402 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.065419 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:32Z","lastTransitionTime":"2026-02-02T15:16:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.081391 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5a81ada-e393-499a-aac6-7b40349958bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4acf2398e2e19bd54f482cb2273c54738e61627187860bb0b6daadddc2565b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831d85055203f153381eb51d98af0cd2e4c4a0309f837545ab3d855a680133e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1380122d1694c1286f1102e6396c6504b91f63c6eb157689dfc0ea7188c17f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5835cb53385359e9c58eaa5d4da2569c313e0fc265ad79a6c124e8cecb1a734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd6dc4bf5b9fd31bfbee6d9340fe981c0c5fca03a79cef9316340ff5a8b3d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db1db75e392ebea8b387b85cabc4276d553bf2c06694cf7f294a383909f32826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55b2c8f5f7c8b659a1b54bb592229f81596804f972fc4658e1f867054721257d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55b2c8f5f7c8b659a1b54bb592229f81596804f972fc4658e1f867054721257d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T15:16:30Z\\\",\\\"message\\\":\\\"lse hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.138:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {97419c58-41c7-41d7-a137-a446f0c7eeb3}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0202 15:16:30.146591 6766 obj_retry.go:551] Creating *factory.egressNode crc took: 3.786809ms\\\\nI0202 15:16:30.147325 6766 factory.go:1336] Added *v1.Node event handler 7\\\\nI0202 15:16:30.147474 6766 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI0202 15:16:30.147563 6766 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-operators\\\\\\\"}\\\\nI0202 15:16:30.147582 6766 services_controller.go:360] Finished syncing service redhat-operators on namespace openshift-marketplace for network=default : 6.641214ms\\\\nI0202 15:16:30.149227 6766 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0202 15:16:30.149857 6766 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0202 15:16:30.150220 6766 ovnkube.go:599] Stopped ovnkube\\\\nI0202 15:16:30.150339 6766 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0202 15:16:30.150729 6766 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:16:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-qxvkf_openshift-ovn-kubernetes(f5a81ada-e393-499a-aac6-7b40349958bb)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efaf2216e4c1750139f080a93b4b608737dab2883de5fddfd9a99248d517b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxvkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:32Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.096885 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8c597" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76cc68c3-9e90-45de-87b2-6ea1b68eadce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ac57831c1c880e64aaa7bedcf999133ee09fc5ecc5124b9d54ecedcc9a13c28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwr7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8c597\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:32Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.113428 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hbm26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45553dc5-54ea-4de9-914d-55b8fee477f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d112b1ea41c3d30a7cfa46d17db6aed0999eb0baaaf8a2f21c741d805a21451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d64j4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f76a1f655f871deb7e1364fcc5dbf6b564c4c386300bb49cb75f83518b64286e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d64j4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hbm26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:32Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.133376 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"402a6dc5-6bc7-4d7c-a909-9bb4ca48365d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d5b032497b592363a03f89cb31ba1b551f479004072dddffce540c41190ab81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a3fd02d55a10647428a05a9f53a1429a9827059c523d5deab3c4581c8f1b933\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11600ada13fbc7db1831de385202bc1f041efb656cc58aa49b0f71fe902db9bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fd39f5a05dec7a756dfd519f5df3ef7c77bbb5b8633c5af18a09ff8029cb079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:32Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.150588 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab21f7f0-7a5d-4eed-bd82-6c9d2e697eda\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ae3c8e412002f52fe150522d9bacb6415308ea41825b90d9d89010a7eca8764\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73fee4346d0af07556e8a1ea336818acea1f39aa8f703056b5954c7fac72aee1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdff2efc8ee6095385be74fbe1bf30926e526639dcc00b39fca8d5c9a0145d3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://742ee25d4d54068f173176593cafc678ec8502b53656fe9c969177f803836e29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://742ee25d4d54068f173176593cafc678ec8502b53656fe9c969177f803836e29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:32Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.168658 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.168708 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.168724 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.168749 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.168768 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:32Z","lastTransitionTime":"2026-02-02T15:16:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.242909 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-24 02:19:48.349700487 +0000 UTC Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.271310 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.272623 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.272673 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.272692 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.272713 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.272732 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:32Z","lastTransitionTime":"2026-02-02T15:16:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.376793 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.376846 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.376862 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.376884 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.376901 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:32Z","lastTransitionTime":"2026-02-02T15:16:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.480598 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.480713 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.480738 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.480768 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.480793 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:32Z","lastTransitionTime":"2026-02-02T15:16:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.584956 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.585026 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.585043 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.585068 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.585087 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:32Z","lastTransitionTime":"2026-02-02T15:16:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.688104 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.688191 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.688211 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.688236 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.688253 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:32Z","lastTransitionTime":"2026-02-02T15:16:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.791696 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.791764 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.791781 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.791806 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.791825 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:32Z","lastTransitionTime":"2026-02-02T15:16:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.895100 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.895204 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.895224 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.895248 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.895266 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:32Z","lastTransitionTime":"2026-02-02T15:16:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.998219 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.998296 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.998314 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.998337 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:32 crc kubenswrapper[4733]: I0202 15:16:32.998354 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:32Z","lastTransitionTime":"2026-02-02T15:16:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:33 crc kubenswrapper[4733]: I0202 15:16:33.101654 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:33 crc kubenswrapper[4733]: I0202 15:16:33.101718 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:33 crc kubenswrapper[4733]: I0202 15:16:33.101735 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:33 crc kubenswrapper[4733]: I0202 15:16:33.101758 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:33 crc kubenswrapper[4733]: I0202 15:16:33.101776 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:33Z","lastTransitionTime":"2026-02-02T15:16:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:33 crc kubenswrapper[4733]: I0202 15:16:33.204822 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:33 crc kubenswrapper[4733]: I0202 15:16:33.204881 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:33 crc kubenswrapper[4733]: I0202 15:16:33.204900 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:33 crc kubenswrapper[4733]: I0202 15:16:33.204923 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:33 crc kubenswrapper[4733]: I0202 15:16:33.204943 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:33Z","lastTransitionTime":"2026-02-02T15:16:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:33 crc kubenswrapper[4733]: I0202 15:16:33.243713 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-21 10:25:55.195228979 +0000 UTC Feb 02 15:16:33 crc kubenswrapper[4733]: I0202 15:16:33.254057 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:16:33 crc kubenswrapper[4733]: I0202 15:16:33.254098 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:16:33 crc kubenswrapper[4733]: I0202 15:16:33.254195 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:16:33 crc kubenswrapper[4733]: E0202 15:16:33.254324 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:16:33 crc kubenswrapper[4733]: I0202 15:16:33.254346 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:16:33 crc kubenswrapper[4733]: E0202 15:16:33.254460 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:16:33 crc kubenswrapper[4733]: E0202 15:16:33.254636 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:16:33 crc kubenswrapper[4733]: E0202 15:16:33.254818 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:16:33 crc kubenswrapper[4733]: I0202 15:16:33.307847 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:33 crc kubenswrapper[4733]: I0202 15:16:33.307905 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:33 crc kubenswrapper[4733]: I0202 15:16:33.307928 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:33 crc kubenswrapper[4733]: I0202 15:16:33.307958 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:33 crc kubenswrapper[4733]: I0202 15:16:33.307982 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:33Z","lastTransitionTime":"2026-02-02T15:16:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:33 crc kubenswrapper[4733]: I0202 15:16:33.411468 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:33 crc kubenswrapper[4733]: I0202 15:16:33.411497 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:33 crc kubenswrapper[4733]: I0202 15:16:33.411505 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:33 crc kubenswrapper[4733]: I0202 15:16:33.411519 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:33 crc kubenswrapper[4733]: I0202 15:16:33.411528 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:33Z","lastTransitionTime":"2026-02-02T15:16:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:33 crc kubenswrapper[4733]: I0202 15:16:33.514073 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:33 crc kubenswrapper[4733]: I0202 15:16:33.514132 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:33 crc kubenswrapper[4733]: I0202 15:16:33.514147 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:33 crc kubenswrapper[4733]: I0202 15:16:33.514207 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:33 crc kubenswrapper[4733]: I0202 15:16:33.514224 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:33Z","lastTransitionTime":"2026-02-02T15:16:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:33 crc kubenswrapper[4733]: I0202 15:16:33.617057 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:33 crc kubenswrapper[4733]: I0202 15:16:33.617134 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:33 crc kubenswrapper[4733]: I0202 15:16:33.617157 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:33 crc kubenswrapper[4733]: I0202 15:16:33.617211 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:33 crc kubenswrapper[4733]: I0202 15:16:33.617229 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:33Z","lastTransitionTime":"2026-02-02T15:16:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:33 crc kubenswrapper[4733]: I0202 15:16:33.720371 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:33 crc kubenswrapper[4733]: I0202 15:16:33.720433 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:33 crc kubenswrapper[4733]: I0202 15:16:33.720451 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:33 crc kubenswrapper[4733]: I0202 15:16:33.720474 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:33 crc kubenswrapper[4733]: I0202 15:16:33.720490 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:33Z","lastTransitionTime":"2026-02-02T15:16:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:33 crc kubenswrapper[4733]: I0202 15:16:33.823301 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:33 crc kubenswrapper[4733]: I0202 15:16:33.823437 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:33 crc kubenswrapper[4733]: I0202 15:16:33.823512 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:33 crc kubenswrapper[4733]: I0202 15:16:33.823543 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:33 crc kubenswrapper[4733]: I0202 15:16:33.823633 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:33Z","lastTransitionTime":"2026-02-02T15:16:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:33 crc kubenswrapper[4733]: I0202 15:16:33.927506 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:33 crc kubenswrapper[4733]: I0202 15:16:33.927597 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:33 crc kubenswrapper[4733]: I0202 15:16:33.927615 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:33 crc kubenswrapper[4733]: I0202 15:16:33.927637 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:33 crc kubenswrapper[4733]: I0202 15:16:33.927686 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:33Z","lastTransitionTime":"2026-02-02T15:16:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:34 crc kubenswrapper[4733]: I0202 15:16:34.030695 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:34 crc kubenswrapper[4733]: I0202 15:16:34.030758 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:34 crc kubenswrapper[4733]: I0202 15:16:34.030775 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:34 crc kubenswrapper[4733]: I0202 15:16:34.030800 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:34 crc kubenswrapper[4733]: I0202 15:16:34.030817 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:34Z","lastTransitionTime":"2026-02-02T15:16:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:34 crc kubenswrapper[4733]: I0202 15:16:34.133470 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:34 crc kubenswrapper[4733]: I0202 15:16:34.133558 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:34 crc kubenswrapper[4733]: I0202 15:16:34.133578 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:34 crc kubenswrapper[4733]: I0202 15:16:34.133603 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:34 crc kubenswrapper[4733]: I0202 15:16:34.133622 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:34Z","lastTransitionTime":"2026-02-02T15:16:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:34 crc kubenswrapper[4733]: I0202 15:16:34.236353 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:34 crc kubenswrapper[4733]: I0202 15:16:34.236414 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:34 crc kubenswrapper[4733]: I0202 15:16:34.236433 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:34 crc kubenswrapper[4733]: I0202 15:16:34.236457 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:34 crc kubenswrapper[4733]: I0202 15:16:34.236475 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:34Z","lastTransitionTime":"2026-02-02T15:16:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:34 crc kubenswrapper[4733]: I0202 15:16:34.243949 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-04 14:08:14.554437729 +0000 UTC Feb 02 15:16:34 crc kubenswrapper[4733]: I0202 15:16:34.339152 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:34 crc kubenswrapper[4733]: I0202 15:16:34.339249 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:34 crc kubenswrapper[4733]: I0202 15:16:34.339267 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:34 crc kubenswrapper[4733]: I0202 15:16:34.339293 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:34 crc kubenswrapper[4733]: I0202 15:16:34.339311 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:34Z","lastTransitionTime":"2026-02-02T15:16:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:34 crc kubenswrapper[4733]: I0202 15:16:34.442417 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:34 crc kubenswrapper[4733]: I0202 15:16:34.442480 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:34 crc kubenswrapper[4733]: I0202 15:16:34.442498 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:34 crc kubenswrapper[4733]: I0202 15:16:34.442522 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:34 crc kubenswrapper[4733]: I0202 15:16:34.442541 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:34Z","lastTransitionTime":"2026-02-02T15:16:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:34 crc kubenswrapper[4733]: I0202 15:16:34.545921 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:34 crc kubenswrapper[4733]: I0202 15:16:34.545979 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:34 crc kubenswrapper[4733]: I0202 15:16:34.545995 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:34 crc kubenswrapper[4733]: I0202 15:16:34.546022 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:34 crc kubenswrapper[4733]: I0202 15:16:34.546041 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:34Z","lastTransitionTime":"2026-02-02T15:16:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:34 crc kubenswrapper[4733]: I0202 15:16:34.648562 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:34 crc kubenswrapper[4733]: I0202 15:16:34.648616 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:34 crc kubenswrapper[4733]: I0202 15:16:34.648633 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:34 crc kubenswrapper[4733]: I0202 15:16:34.648654 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:34 crc kubenswrapper[4733]: I0202 15:16:34.648671 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:34Z","lastTransitionTime":"2026-02-02T15:16:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:34 crc kubenswrapper[4733]: I0202 15:16:34.751394 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:34 crc kubenswrapper[4733]: I0202 15:16:34.751479 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:34 crc kubenswrapper[4733]: I0202 15:16:34.751496 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:34 crc kubenswrapper[4733]: I0202 15:16:34.751521 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:34 crc kubenswrapper[4733]: I0202 15:16:34.751539 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:34Z","lastTransitionTime":"2026-02-02T15:16:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:34 crc kubenswrapper[4733]: I0202 15:16:34.855079 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:34 crc kubenswrapper[4733]: I0202 15:16:34.855138 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:34 crc kubenswrapper[4733]: I0202 15:16:34.855156 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:34 crc kubenswrapper[4733]: I0202 15:16:34.855214 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:34 crc kubenswrapper[4733]: I0202 15:16:34.855233 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:34Z","lastTransitionTime":"2026-02-02T15:16:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:34 crc kubenswrapper[4733]: I0202 15:16:34.958322 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:34 crc kubenswrapper[4733]: I0202 15:16:34.958392 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:34 crc kubenswrapper[4733]: I0202 15:16:34.958409 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:34 crc kubenswrapper[4733]: I0202 15:16:34.958432 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:34 crc kubenswrapper[4733]: I0202 15:16:34.958450 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:34Z","lastTransitionTime":"2026-02-02T15:16:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.061598 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.061675 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.061696 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.061726 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.061747 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:35Z","lastTransitionTime":"2026-02-02T15:16:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.164888 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.164943 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.164960 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.164983 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.164999 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:35Z","lastTransitionTime":"2026-02-02T15:16:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.244557 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-14 15:49:19.699076519 +0000 UTC Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.254200 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.254220 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.254344 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.254451 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:16:35 crc kubenswrapper[4733]: E0202 15:16:35.254653 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:16:35 crc kubenswrapper[4733]: E0202 15:16:35.254776 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:16:35 crc kubenswrapper[4733]: E0202 15:16:35.254932 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:16:35 crc kubenswrapper[4733]: E0202 15:16:35.255233 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.267687 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.267743 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.267760 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.267782 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.267800 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:35Z","lastTransitionTime":"2026-02-02T15:16:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.371955 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.372016 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.372032 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.372056 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.372076 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:35Z","lastTransitionTime":"2026-02-02T15:16:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.475274 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.475362 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.475381 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.475404 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.475452 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:35Z","lastTransitionTime":"2026-02-02T15:16:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.579156 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.579263 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.579281 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.579306 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.579322 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:35Z","lastTransitionTime":"2026-02-02T15:16:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.682524 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.682587 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.682604 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.682628 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.682645 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:35Z","lastTransitionTime":"2026-02-02T15:16:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.785089 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.785159 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.785226 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.785259 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.785280 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:35Z","lastTransitionTime":"2026-02-02T15:16:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.887655 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.887715 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.887733 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.887755 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.887774 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:35Z","lastTransitionTime":"2026-02-02T15:16:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.990927 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.990972 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.990983 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.991000 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:35 crc kubenswrapper[4733]: I0202 15:16:35.991011 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:35Z","lastTransitionTime":"2026-02-02T15:16:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:36 crc kubenswrapper[4733]: I0202 15:16:36.093898 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:36 crc kubenswrapper[4733]: I0202 15:16:36.093966 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:36 crc kubenswrapper[4733]: I0202 15:16:36.093987 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:36 crc kubenswrapper[4733]: I0202 15:16:36.094011 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:36 crc kubenswrapper[4733]: I0202 15:16:36.094028 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:36Z","lastTransitionTime":"2026-02-02T15:16:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:36 crc kubenswrapper[4733]: I0202 15:16:36.197700 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:36 crc kubenswrapper[4733]: I0202 15:16:36.197744 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:36 crc kubenswrapper[4733]: I0202 15:16:36.197762 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:36 crc kubenswrapper[4733]: I0202 15:16:36.197785 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:36 crc kubenswrapper[4733]: I0202 15:16:36.197803 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:36Z","lastTransitionTime":"2026-02-02T15:16:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:36 crc kubenswrapper[4733]: I0202 15:16:36.245387 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-02 07:03:12.116587277 +0000 UTC Feb 02 15:16:36 crc kubenswrapper[4733]: I0202 15:16:36.302445 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:36 crc kubenswrapper[4733]: I0202 15:16:36.302798 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:36 crc kubenswrapper[4733]: I0202 15:16:36.302969 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:36 crc kubenswrapper[4733]: I0202 15:16:36.303488 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:36 crc kubenswrapper[4733]: I0202 15:16:36.303642 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:36Z","lastTransitionTime":"2026-02-02T15:16:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:36 crc kubenswrapper[4733]: I0202 15:16:36.406955 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:36 crc kubenswrapper[4733]: I0202 15:16:36.407373 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:36 crc kubenswrapper[4733]: I0202 15:16:36.407559 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:36 crc kubenswrapper[4733]: I0202 15:16:36.407732 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:36 crc kubenswrapper[4733]: I0202 15:16:36.407886 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:36Z","lastTransitionTime":"2026-02-02T15:16:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:36 crc kubenswrapper[4733]: I0202 15:16:36.511450 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:36 crc kubenswrapper[4733]: I0202 15:16:36.511535 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:36 crc kubenswrapper[4733]: I0202 15:16:36.511558 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:36 crc kubenswrapper[4733]: I0202 15:16:36.511590 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:36 crc kubenswrapper[4733]: I0202 15:16:36.511612 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:36Z","lastTransitionTime":"2026-02-02T15:16:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:36 crc kubenswrapper[4733]: I0202 15:16:36.614651 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:36 crc kubenswrapper[4733]: I0202 15:16:36.614703 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:36 crc kubenswrapper[4733]: I0202 15:16:36.614721 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:36 crc kubenswrapper[4733]: I0202 15:16:36.614747 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:36 crc kubenswrapper[4733]: I0202 15:16:36.614765 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:36Z","lastTransitionTime":"2026-02-02T15:16:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:36 crc kubenswrapper[4733]: I0202 15:16:36.717921 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:36 crc kubenswrapper[4733]: I0202 15:16:36.718064 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:36 crc kubenswrapper[4733]: I0202 15:16:36.718088 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:36 crc kubenswrapper[4733]: I0202 15:16:36.718119 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:36 crc kubenswrapper[4733]: I0202 15:16:36.718141 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:36Z","lastTransitionTime":"2026-02-02T15:16:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:36 crc kubenswrapper[4733]: I0202 15:16:36.821550 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:36 crc kubenswrapper[4733]: I0202 15:16:36.821616 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:36 crc kubenswrapper[4733]: I0202 15:16:36.821637 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:36 crc kubenswrapper[4733]: I0202 15:16:36.821661 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:36 crc kubenswrapper[4733]: I0202 15:16:36.821679 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:36Z","lastTransitionTime":"2026-02-02T15:16:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:36 crc kubenswrapper[4733]: I0202 15:16:36.925052 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:36 crc kubenswrapper[4733]: I0202 15:16:36.925431 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:36 crc kubenswrapper[4733]: I0202 15:16:36.925547 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:36 crc kubenswrapper[4733]: I0202 15:16:36.925645 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:36 crc kubenswrapper[4733]: I0202 15:16:36.925724 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:36Z","lastTransitionTime":"2026-02-02T15:16:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.029412 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.029529 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.029554 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.029585 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.029606 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:37Z","lastTransitionTime":"2026-02-02T15:16:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.132896 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.132960 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.132980 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.133007 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.133028 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:37Z","lastTransitionTime":"2026-02-02T15:16:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.235993 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.236043 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.236059 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.236085 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.236107 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:37Z","lastTransitionTime":"2026-02-02T15:16:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.246342 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-10 22:26:18.745587151 +0000 UTC Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.254682 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.254826 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:16:37 crc kubenswrapper[4733]: E0202 15:16:37.255003 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.255054 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.255112 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:16:37 crc kubenswrapper[4733]: E0202 15:16:37.255354 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:16:37 crc kubenswrapper[4733]: E0202 15:16:37.255467 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:16:37 crc kubenswrapper[4733]: E0202 15:16:37.255561 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.275806 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca57dabf1d1ee5c6fc29291b1be687ca7e97c72650084b02067057fccac41a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:37Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.297015 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:37Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.312794 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5xgpn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56681968-f264-440e-aa12-d4f81879efe9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e250891b1065a38a448aa6abd300b37543e819e87f006b03cc37983cb643982a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdjjl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5xgpn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:37Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.327100 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-7r9fd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4dfd44b-b363-4162-9f43-9a8dc4a3b110\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jtscj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-7r9fd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:37Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.340086 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.340206 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.340233 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.340315 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.340338 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:37Z","lastTransitionTime":"2026-02-02T15:16:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.353207 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7082da2-f2e7-4b73-ba35-504e5cd4c920\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd383d69b9b27d52fa8672ff6e0294a74876d676973bdc0ca95125fef5f2d085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037e379a109ce230098e3ca223fd1dea3149de3ebe601f56e2bb295f3093a3c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6fdd19dac506f2d5d7eedc44ec2cda92a9381b73e91a45d949e4fcde4dfc283\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13965f5abcfb0a36c33310518955f192673e630a1aaba94b49400026de5bc2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://992f4198192cbe4b1c89d575aa4a20c29c1115fd5eaa8986c6ed70c1d6b143c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-02T15:15:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0202 15:15:31.010033 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0202 15:15:31.011243 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3543528698/tls.crt::/tmp/serving-cert-3543528698/tls.key\\\\\\\"\\\\nI0202 15:15:36.416658 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0202 15:15:36.428087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0202 15:15:36.428134 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0202 15:15:36.428243 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0202 15:15:36.428280 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0202 15:15:36.439678 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0202 15:15:36.439710 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439717 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0202 15:15:36.439722 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0202 15:15:36.439726 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0202 15:15:36.439729 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0202 15:15:36.439732 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0202 15:15:36.439753 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0202 15:15:36.450716 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://226f5949820d4f498dbf6fa251c6edf1c6356900bd946e6afae2c29042b1c835\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:37Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.373982 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44d77bd7-65fb-447b-b806-c096251a5d69\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c25f4250ab2773e407dd24b8b81e0f4bf800d2cf3f7d5c3f50ffbd7f4c755c95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e8b7ecb2cfad9071bb3cc7968db64e99dc2fc24df7833a21ef302e062b5d3ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e8b7ecb2cfad9071bb3cc7968db64e99dc2fc24df7833a21ef302e062b5d3ce\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:37Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.396109 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da08f9aecc070b31df2ddc1be7c6eb2ad8ef68a931e57ca2e5c5302f4e75751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:37Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.416820 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:37Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.435006 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hbm26" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45553dc5-54ea-4de9-914d-55b8fee477f8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d112b1ea41c3d30a7cfa46d17db6aed0999eb0baaaf8a2f21c741d805a21451\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d64j4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f76a1f655f871deb7e1364fcc5dbf6b564c4c386300bb49cb75f83518b64286e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d64j4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hbm26\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:37Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.443525 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.443594 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.443619 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.443649 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.443671 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:37Z","lastTransitionTime":"2026-02-02T15:16:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.469121 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8206eca4-7d8b-4b59-a4c1-24631dd8cd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41d4a7bdc21f00b5039385d7227d8cb0f4ea3051d3b5065c71a49cb0bfad26a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c11b1859bf3bc0ec4d5e6a58d9becc1ac6333a1ed9820d83d634f1e93221c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c08b087d85be9bb83007c7e38a5de9182a5091716fb8a2ca5ded1b8191bebc40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fd4969f0f16c9aa44208d2bf578f77d9d50914a8cb1764507175aa2eb58e733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59175c8e5a7f3536168a69fd944bf3d536bcd7281eee96002343b7836a54401b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dde1061684c05edb8148e5c5c5679768ddbcf78b71239ee1c801ecdd100fc71\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc2989d1fc99a58f060258626160d78c6786a1b5ace57fbcccbcbc63079fbfcf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0e66fccbbdabed0005722703fc3c078762f4d3eb43dc51a7e38b391316996e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:37Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.493662 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nr6ps" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e43cd902-791a-4042-bc08-cf03e075be32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2a89a5d1fc43696f33d313aee5dca316a249fdef6a23ff655e31f14cc449cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8f4fea86ef4f622881b9a35cbc0e6f3003be0937b4b564fe6ff101235ec018c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8454f16b88cf4d04b7197caaaeaa14ba2bd09fef9e6389516a6033acb6261cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://019267f23fe7a3e84e220a2173d499be047719b68a5f23361c6c8b195245a347\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52f7783b5dc28fb31c99fb904af52493e1de60dacdc453ba360d9b2ccd978722\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07e56782051c93b2a320aa539a42764b97647237d228bed2b6af48cee987c60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07e56782051c93b2a320aa539a42764b97647237d228bed2b6af48cee987c60b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce850ac7bd4c2444b65d87232114d55b78d11c6f973537c615900092a45d55e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce850ac7bd4c2444b65d87232114d55b78d11c6f973537c615900092a45d55e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mfskw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nr6ps\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:37Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.525783 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f5a81ada-e393-499a-aac6-7b40349958bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4acf2398e2e19bd54f482cb2273c54738e61627187860bb0b6daadddc2565b63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://831d85055203f153381eb51d98af0cd2e4c4a0309f837545ab3d855a680133e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1380122d1694c1286f1102e6396c6504b91f63c6eb157689dfc0ea7188c17f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5835cb53385359e9c58eaa5d4da2569c313e0fc265ad79a6c124e8cecb1a734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffd6dc4bf5b9fd31bfbee6d9340fe981c0c5fca03a79cef9316340ff5a8b3d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db1db75e392ebea8b387b85cabc4276d553bf2c06694cf7f294a383909f32826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55b2c8f5f7c8b659a1b54bb592229f81596804f972fc4658e1f867054721257d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55b2c8f5f7c8b659a1b54bb592229f81596804f972fc4658e1f867054721257d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T15:16:30Z\\\",\\\"message\\\":\\\"lse hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.138:50051:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {97419c58-41c7-41d7-a137-a446f0c7eeb3}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0202 15:16:30.146591 6766 obj_retry.go:551] Creating *factory.egressNode crc took: 3.786809ms\\\\nI0202 15:16:30.147325 6766 factory.go:1336] Added *v1.Node event handler 7\\\\nI0202 15:16:30.147474 6766 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI0202 15:16:30.147563 6766 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-operators\\\\\\\"}\\\\nI0202 15:16:30.147582 6766 services_controller.go:360] Finished syncing service redhat-operators on namespace openshift-marketplace for network=default : 6.641214ms\\\\nI0202 15:16:30.149227 6766 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0202 15:16:30.149857 6766 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0202 15:16:30.150220 6766 ovnkube.go:599] Stopped ovnkube\\\\nI0202 15:16:30.150339 6766 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0202 15:16:30.150729 6766 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:16:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-qxvkf_openshift-ovn-kubernetes(f5a81ada-e393-499a-aac6-7b40349958bb)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6efaf2216e4c1750139f080a93b4b608737dab2883de5fddfd9a99248d517b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4z57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxvkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:37Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.540042 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8c597" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"76cc68c3-9e90-45de-87b2-6ea1b68eadce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ac57831c1c880e64aaa7bedcf999133ee09fc5ecc5124b9d54ecedcc9a13c28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qwr7v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8c597\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:37Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.549876 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.549915 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.549931 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.549945 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.549955 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:37Z","lastTransitionTime":"2026-02-02T15:16:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.568237 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"402a6dc5-6bc7-4d7c-a909-9bb4ca48365d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d5b032497b592363a03f89cb31ba1b551f479004072dddffce540c41190ab81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a3fd02d55a10647428a05a9f53a1429a9827059c523d5deab3c4581c8f1b933\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11600ada13fbc7db1831de385202bc1f041efb656cc58aa49b0f71fe902db9bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8fd39f5a05dec7a756dfd519f5df3ef7c77bbb5b8633c5af18a09ff8029cb079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:37Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.585103 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab21f7f0-7a5d-4eed-bd82-6c9d2e697eda\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ae3c8e412002f52fe150522d9bacb6415308ea41825b90d9d89010a7eca8764\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73fee4346d0af07556e8a1ea336818acea1f39aa8f703056b5954c7fac72aee1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdff2efc8ee6095385be74fbe1bf30926e526639dcc00b39fca8d5c9a0145d3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://742ee25d4d54068f173176593cafc678ec8502b53656fe9c969177f803836e29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://742ee25d4d54068f173176593cafc678ec8502b53656fe9c969177f803836e29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-02T15:15:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:37Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.607747 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:37Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.628082 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d65b38b8410bea95465ee58803c4160039bf7c082ca47893680da7469418f429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27003b336112760daa2079f4819641850c27fbc8415b63b5094c86195291aca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:37Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.653157 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.653285 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.653305 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.653121 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jh6vk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76a05d6-d9c7-4183-ad90-b3131f225841\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f57aea989227fdbd7bb59c51fade50d1076929f6288d060808b51bd1bdd7df04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f2659e2a914d3191daddbfb5ad109e487f03dabb559d324244e127add7d41d9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-02T15:16:24Z\\\",\\\"message\\\":\\\"2026-02-02T15:15:39+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_bbf1a3b4-24d2-4558-bc15-4414386de7b4\\\\n2026-02-02T15:15:39+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_bbf1a3b4-24d2-4558-bc15-4414386de7b4 to /host/opt/cni/bin/\\\\n2026-02-02T15:15:39Z [verbose] multus-daemon started\\\\n2026-02-02T15:15:39Z [verbose] Readiness Indicator file check\\\\n2026-02-02T15:16:24Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hv9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jh6vk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:37Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.653329 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.653575 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:37Z","lastTransitionTime":"2026-02-02T15:16:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.677269 4733 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"83d0e810-ea5c-47aa-947b-7865b045bf94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-02T15:15:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6144eb32f88bc106f7724e48133b568a137c4f3fbbfe30c7fcef405d3ab735a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51318ad71b42f4cc8d376aa50e44580e68da89d83d80defb996329a2ee5c2ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-02T15:15:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hz28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-02T15:15:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zgsv9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:37Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.757001 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.757123 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.757230 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.757307 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.757333 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:37Z","lastTransitionTime":"2026-02-02T15:16:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.859914 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.859954 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.859965 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.859981 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.859992 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:37Z","lastTransitionTime":"2026-02-02T15:16:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.962687 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.962758 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.962776 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.962800 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:37 crc kubenswrapper[4733]: I0202 15:16:37.962817 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:37Z","lastTransitionTime":"2026-02-02T15:16:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.001105 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.001207 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.001228 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.001253 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.001271 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:38Z","lastTransitionTime":"2026-02-02T15:16:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:38 crc kubenswrapper[4733]: E0202 15:16:38.022354 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd742269-802e-45ec-a88e-17aee17c574e\\\",\\\"systemUUID\\\":\\\"91b64b75-3cd9-4fb9-93f0-dbd3bcdcaec4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:38Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.027635 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.027736 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.027757 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.027817 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.027837 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:38Z","lastTransitionTime":"2026-02-02T15:16:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:38 crc kubenswrapper[4733]: E0202 15:16:38.050455 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd742269-802e-45ec-a88e-17aee17c574e\\\",\\\"systemUUID\\\":\\\"91b64b75-3cd9-4fb9-93f0-dbd3bcdcaec4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:38Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.056872 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.056938 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.056956 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.056984 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.057006 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:38Z","lastTransitionTime":"2026-02-02T15:16:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:38 crc kubenswrapper[4733]: E0202 15:16:38.074085 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd742269-802e-45ec-a88e-17aee17c574e\\\",\\\"systemUUID\\\":\\\"91b64b75-3cd9-4fb9-93f0-dbd3bcdcaec4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:38Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.079772 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.080261 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.080283 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.080307 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.080328 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:38Z","lastTransitionTime":"2026-02-02T15:16:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:38 crc kubenswrapper[4733]: E0202 15:16:38.103241 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd742269-802e-45ec-a88e-17aee17c574e\\\",\\\"systemUUID\\\":\\\"91b64b75-3cd9-4fb9-93f0-dbd3bcdcaec4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:38Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.108490 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.108537 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.108553 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.108574 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.108591 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:38Z","lastTransitionTime":"2026-02-02T15:16:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:38 crc kubenswrapper[4733]: E0202 15:16:38.129771 4733 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-02T15:16:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-02T15:16:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cd742269-802e-45ec-a88e-17aee17c574e\\\",\\\"systemUUID\\\":\\\"91b64b75-3cd9-4fb9-93f0-dbd3bcdcaec4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-02T15:16:38Z is after 2025-08-24T17:21:41Z" Feb 02 15:16:38 crc kubenswrapper[4733]: E0202 15:16:38.129989 4733 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.132106 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.132157 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.132221 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.132243 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.132260 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:38Z","lastTransitionTime":"2026-02-02T15:16:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.235054 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.235137 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.235196 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.235226 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.235256 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:38Z","lastTransitionTime":"2026-02-02T15:16:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.247238 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-04 17:15:11.718296513 +0000 UTC Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.338476 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.338563 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.338595 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.338624 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.338644 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:38Z","lastTransitionTime":"2026-02-02T15:16:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.440745 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.440824 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.440846 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.440877 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.440899 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:38Z","lastTransitionTime":"2026-02-02T15:16:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.543640 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.543720 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.543777 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.543822 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.543843 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:38Z","lastTransitionTime":"2026-02-02T15:16:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.646697 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.646759 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.646777 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.646800 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.646817 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:38Z","lastTransitionTime":"2026-02-02T15:16:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.750710 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.750774 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.750791 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.750817 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.750835 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:38Z","lastTransitionTime":"2026-02-02T15:16:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.858745 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.858810 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.858828 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.858853 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.858870 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:38Z","lastTransitionTime":"2026-02-02T15:16:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.961590 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.961644 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.961659 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.961681 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:38 crc kubenswrapper[4733]: I0202 15:16:38.961696 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:38Z","lastTransitionTime":"2026-02-02T15:16:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.064871 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.064943 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.064960 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.064985 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.065002 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:39Z","lastTransitionTime":"2026-02-02T15:16:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.168198 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.168264 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.168282 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.168309 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.168326 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:39Z","lastTransitionTime":"2026-02-02T15:16:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.247363 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 22:15:04.886902868 +0000 UTC Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.254808 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.254839 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:16:39 crc kubenswrapper[4733]: E0202 15:16:39.254998 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.255059 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.255081 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:16:39 crc kubenswrapper[4733]: E0202 15:16:39.255285 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:16:39 crc kubenswrapper[4733]: E0202 15:16:39.255406 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:16:39 crc kubenswrapper[4733]: E0202 15:16:39.255498 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.271251 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.271283 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.271294 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.271307 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.271321 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:39Z","lastTransitionTime":"2026-02-02T15:16:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.374434 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.374476 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.374487 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.374502 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.374513 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:39Z","lastTransitionTime":"2026-02-02T15:16:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.476763 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.476853 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.476872 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.476931 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.476951 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:39Z","lastTransitionTime":"2026-02-02T15:16:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.579620 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.579682 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.579702 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.579729 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.579750 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:39Z","lastTransitionTime":"2026-02-02T15:16:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.682475 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.682526 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.682544 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.682570 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.682588 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:39Z","lastTransitionTime":"2026-02-02T15:16:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.785551 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.785916 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.786076 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.786289 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.786437 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:39Z","lastTransitionTime":"2026-02-02T15:16:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.889378 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.889414 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.889424 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.889438 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.889448 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:39Z","lastTransitionTime":"2026-02-02T15:16:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.993024 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.993079 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.993095 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.993128 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:39 crc kubenswrapper[4733]: I0202 15:16:39.993145 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:39Z","lastTransitionTime":"2026-02-02T15:16:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:40 crc kubenswrapper[4733]: I0202 15:16:40.095700 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:40 crc kubenswrapper[4733]: I0202 15:16:40.096106 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:40 crc kubenswrapper[4733]: I0202 15:16:40.096389 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:40 crc kubenswrapper[4733]: I0202 15:16:40.096760 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:40 crc kubenswrapper[4733]: I0202 15:16:40.097299 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:40Z","lastTransitionTime":"2026-02-02T15:16:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:40 crc kubenswrapper[4733]: I0202 15:16:40.200254 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:40 crc kubenswrapper[4733]: I0202 15:16:40.200331 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:40 crc kubenswrapper[4733]: I0202 15:16:40.200357 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:40 crc kubenswrapper[4733]: I0202 15:16:40.200388 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:40 crc kubenswrapper[4733]: I0202 15:16:40.200448 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:40Z","lastTransitionTime":"2026-02-02T15:16:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:40 crc kubenswrapper[4733]: I0202 15:16:40.248315 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-03 06:56:33.992921838 +0000 UTC Feb 02 15:16:40 crc kubenswrapper[4733]: I0202 15:16:40.302940 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:40 crc kubenswrapper[4733]: I0202 15:16:40.302991 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:40 crc kubenswrapper[4733]: I0202 15:16:40.303008 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:40 crc kubenswrapper[4733]: I0202 15:16:40.303030 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:40 crc kubenswrapper[4733]: I0202 15:16:40.303050 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:40Z","lastTransitionTime":"2026-02-02T15:16:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:40 crc kubenswrapper[4733]: I0202 15:16:40.406352 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:40 crc kubenswrapper[4733]: I0202 15:16:40.406416 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:40 crc kubenswrapper[4733]: I0202 15:16:40.406437 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:40 crc kubenswrapper[4733]: I0202 15:16:40.406465 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:40 crc kubenswrapper[4733]: I0202 15:16:40.406484 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:40Z","lastTransitionTime":"2026-02-02T15:16:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:40 crc kubenswrapper[4733]: I0202 15:16:40.510289 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:40 crc kubenswrapper[4733]: I0202 15:16:40.510368 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:40 crc kubenswrapper[4733]: I0202 15:16:40.510392 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:40 crc kubenswrapper[4733]: I0202 15:16:40.510423 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:40 crc kubenswrapper[4733]: I0202 15:16:40.510445 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:40Z","lastTransitionTime":"2026-02-02T15:16:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:40 crc kubenswrapper[4733]: I0202 15:16:40.613781 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:40 crc kubenswrapper[4733]: I0202 15:16:40.613841 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:40 crc kubenswrapper[4733]: I0202 15:16:40.613858 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:40 crc kubenswrapper[4733]: I0202 15:16:40.613885 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:40 crc kubenswrapper[4733]: I0202 15:16:40.613903 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:40Z","lastTransitionTime":"2026-02-02T15:16:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:40 crc kubenswrapper[4733]: I0202 15:16:40.716479 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:40 crc kubenswrapper[4733]: I0202 15:16:40.716528 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:40 crc kubenswrapper[4733]: I0202 15:16:40.716540 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:40 crc kubenswrapper[4733]: I0202 15:16:40.716556 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:40 crc kubenswrapper[4733]: I0202 15:16:40.716569 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:40Z","lastTransitionTime":"2026-02-02T15:16:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:40 crc kubenswrapper[4733]: I0202 15:16:40.819480 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:40 crc kubenswrapper[4733]: I0202 15:16:40.819536 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:40 crc kubenswrapper[4733]: I0202 15:16:40.819548 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:40 crc kubenswrapper[4733]: I0202 15:16:40.819564 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:40 crc kubenswrapper[4733]: I0202 15:16:40.819944 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:40Z","lastTransitionTime":"2026-02-02T15:16:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:40 crc kubenswrapper[4733]: I0202 15:16:40.922715 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:40 crc kubenswrapper[4733]: I0202 15:16:40.922801 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:40 crc kubenswrapper[4733]: I0202 15:16:40.922822 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:40 crc kubenswrapper[4733]: I0202 15:16:40.922843 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:40 crc kubenswrapper[4733]: I0202 15:16:40.922899 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:40Z","lastTransitionTime":"2026-02-02T15:16:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.013668 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 15:16:41 crc kubenswrapper[4733]: E0202 15:16:41.014000 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 15:17:45.013961752 +0000 UTC m=+148.465423150 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.014127 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:16:41 crc kubenswrapper[4733]: E0202 15:16:41.014324 4733 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 02 15:16:41 crc kubenswrapper[4733]: E0202 15:16:41.014399 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-02 15:17:45.014382433 +0000 UTC m=+148.465843831 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.025613 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.025652 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.025663 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.025679 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.025690 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:41Z","lastTransitionTime":"2026-02-02T15:16:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.115376 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d4dfd44b-b363-4162-9f43-9a8dc4a3b110-metrics-certs\") pod \"network-metrics-daemon-7r9fd\" (UID: \"d4dfd44b-b363-4162-9f43-9a8dc4a3b110\") " pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.115443 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.115480 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.115524 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:16:41 crc kubenswrapper[4733]: E0202 15:16:41.115649 4733 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 02 15:16:41 crc kubenswrapper[4733]: E0202 15:16:41.115720 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-02 15:17:45.115697962 +0000 UTC m=+148.567159360 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 02 15:16:41 crc kubenswrapper[4733]: E0202 15:16:41.115821 4733 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 02 15:16:41 crc kubenswrapper[4733]: E0202 15:16:41.115892 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d4dfd44b-b363-4162-9f43-9a8dc4a3b110-metrics-certs podName:d4dfd44b-b363-4162-9f43-9a8dc4a3b110 nodeName:}" failed. No retries permitted until 2026-02-02 15:17:45.115864156 +0000 UTC m=+148.567325584 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d4dfd44b-b363-4162-9f43-9a8dc4a3b110-metrics-certs") pod "network-metrics-daemon-7r9fd" (UID: "d4dfd44b-b363-4162-9f43-9a8dc4a3b110") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 02 15:16:41 crc kubenswrapper[4733]: E0202 15:16:41.116005 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 02 15:16:41 crc kubenswrapper[4733]: E0202 15:16:41.116027 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 02 15:16:41 crc kubenswrapper[4733]: E0202 15:16:41.116046 4733 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 15:16:41 crc kubenswrapper[4733]: E0202 15:16:41.116094 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-02 15:17:45.116077602 +0000 UTC m=+148.567539000 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 15:16:41 crc kubenswrapper[4733]: E0202 15:16:41.116227 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 02 15:16:41 crc kubenswrapper[4733]: E0202 15:16:41.116284 4733 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 02 15:16:41 crc kubenswrapper[4733]: E0202 15:16:41.116305 4733 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 15:16:41 crc kubenswrapper[4733]: E0202 15:16:41.116400 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-02 15:17:45.116373909 +0000 UTC m=+148.567835307 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.128388 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.128433 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.128450 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.128472 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.128489 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:41Z","lastTransitionTime":"2026-02-02T15:16:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.231988 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.232048 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.232066 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.232090 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.232108 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:41Z","lastTransitionTime":"2026-02-02T15:16:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.249617 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 01:22:01.803127544 +0000 UTC Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.254065 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.254119 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:16:41 crc kubenswrapper[4733]: E0202 15:16:41.254303 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.254343 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.254443 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:16:41 crc kubenswrapper[4733]: E0202 15:16:41.254594 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:16:41 crc kubenswrapper[4733]: E0202 15:16:41.254786 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:16:41 crc kubenswrapper[4733]: E0202 15:16:41.254906 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.335039 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.335139 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.335215 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.335251 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.335308 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:41Z","lastTransitionTime":"2026-02-02T15:16:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.438723 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.438796 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.438819 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.438848 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.438870 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:41Z","lastTransitionTime":"2026-02-02T15:16:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.542344 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.542398 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.542414 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.542438 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.542460 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:41Z","lastTransitionTime":"2026-02-02T15:16:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.645392 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.645506 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.645529 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.645559 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.645580 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:41Z","lastTransitionTime":"2026-02-02T15:16:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.749463 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.749520 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.749537 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.749559 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.749589 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:41Z","lastTransitionTime":"2026-02-02T15:16:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.851895 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.851957 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.851973 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.851997 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.852014 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:41Z","lastTransitionTime":"2026-02-02T15:16:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.954352 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.954518 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.954535 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.954556 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:41 crc kubenswrapper[4733]: I0202 15:16:41.954571 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:41Z","lastTransitionTime":"2026-02-02T15:16:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:42 crc kubenswrapper[4733]: I0202 15:16:42.057277 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:42 crc kubenswrapper[4733]: I0202 15:16:42.057347 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:42 crc kubenswrapper[4733]: I0202 15:16:42.057364 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:42 crc kubenswrapper[4733]: I0202 15:16:42.057388 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:42 crc kubenswrapper[4733]: I0202 15:16:42.057408 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:42Z","lastTransitionTime":"2026-02-02T15:16:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:42 crc kubenswrapper[4733]: I0202 15:16:42.160133 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:42 crc kubenswrapper[4733]: I0202 15:16:42.160227 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:42 crc kubenswrapper[4733]: I0202 15:16:42.160245 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:42 crc kubenswrapper[4733]: I0202 15:16:42.160269 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:42 crc kubenswrapper[4733]: I0202 15:16:42.160289 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:42Z","lastTransitionTime":"2026-02-02T15:16:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:42 crc kubenswrapper[4733]: I0202 15:16:42.250324 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 05:24:04.747246594 +0000 UTC Feb 02 15:16:42 crc kubenswrapper[4733]: I0202 15:16:42.256041 4733 scope.go:117] "RemoveContainer" containerID="55b2c8f5f7c8b659a1b54bb592229f81596804f972fc4658e1f867054721257d" Feb 02 15:16:42 crc kubenswrapper[4733]: E0202 15:16:42.256345 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-qxvkf_openshift-ovn-kubernetes(f5a81ada-e393-499a-aac6-7b40349958bb)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" Feb 02 15:16:42 crc kubenswrapper[4733]: I0202 15:16:42.262917 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:42 crc kubenswrapper[4733]: I0202 15:16:42.263001 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:42 crc kubenswrapper[4733]: I0202 15:16:42.263021 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:42 crc kubenswrapper[4733]: I0202 15:16:42.263048 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:42 crc kubenswrapper[4733]: I0202 15:16:42.263065 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:42Z","lastTransitionTime":"2026-02-02T15:16:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:42 crc kubenswrapper[4733]: I0202 15:16:42.364822 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:42 crc kubenswrapper[4733]: I0202 15:16:42.364944 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:42 crc kubenswrapper[4733]: I0202 15:16:42.364963 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:42 crc kubenswrapper[4733]: I0202 15:16:42.364987 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:42 crc kubenswrapper[4733]: I0202 15:16:42.365004 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:42Z","lastTransitionTime":"2026-02-02T15:16:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:42 crc kubenswrapper[4733]: I0202 15:16:42.468251 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:42 crc kubenswrapper[4733]: I0202 15:16:42.468305 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:42 crc kubenswrapper[4733]: I0202 15:16:42.468318 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:42 crc kubenswrapper[4733]: I0202 15:16:42.468339 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:42 crc kubenswrapper[4733]: I0202 15:16:42.468355 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:42Z","lastTransitionTime":"2026-02-02T15:16:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:42 crc kubenswrapper[4733]: I0202 15:16:42.571806 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:42 crc kubenswrapper[4733]: I0202 15:16:42.571869 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:42 crc kubenswrapper[4733]: I0202 15:16:42.571886 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:42 crc kubenswrapper[4733]: I0202 15:16:42.571909 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:42 crc kubenswrapper[4733]: I0202 15:16:42.571931 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:42Z","lastTransitionTime":"2026-02-02T15:16:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:42 crc kubenswrapper[4733]: I0202 15:16:42.674683 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:42 crc kubenswrapper[4733]: I0202 15:16:42.674745 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:42 crc kubenswrapper[4733]: I0202 15:16:42.674762 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:42 crc kubenswrapper[4733]: I0202 15:16:42.674785 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:42 crc kubenswrapper[4733]: I0202 15:16:42.674802 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:42Z","lastTransitionTime":"2026-02-02T15:16:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:42 crc kubenswrapper[4733]: I0202 15:16:42.777787 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:42 crc kubenswrapper[4733]: I0202 15:16:42.777839 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:42 crc kubenswrapper[4733]: I0202 15:16:42.777860 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:42 crc kubenswrapper[4733]: I0202 15:16:42.777889 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:42 crc kubenswrapper[4733]: I0202 15:16:42.777911 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:42Z","lastTransitionTime":"2026-02-02T15:16:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:42 crc kubenswrapper[4733]: I0202 15:16:42.881217 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:42 crc kubenswrapper[4733]: I0202 15:16:42.881266 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:42 crc kubenswrapper[4733]: I0202 15:16:42.881281 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:42 crc kubenswrapper[4733]: I0202 15:16:42.881304 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:42 crc kubenswrapper[4733]: I0202 15:16:42.881319 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:42Z","lastTransitionTime":"2026-02-02T15:16:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:42 crc kubenswrapper[4733]: I0202 15:16:42.984352 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:42 crc kubenswrapper[4733]: I0202 15:16:42.984417 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:42 crc kubenswrapper[4733]: I0202 15:16:42.984439 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:42 crc kubenswrapper[4733]: I0202 15:16:42.984473 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:42 crc kubenswrapper[4733]: I0202 15:16:42.984495 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:42Z","lastTransitionTime":"2026-02-02T15:16:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:43 crc kubenswrapper[4733]: I0202 15:16:43.087254 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:43 crc kubenswrapper[4733]: I0202 15:16:43.087313 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:43 crc kubenswrapper[4733]: I0202 15:16:43.087330 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:43 crc kubenswrapper[4733]: I0202 15:16:43.087359 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:43 crc kubenswrapper[4733]: I0202 15:16:43.087377 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:43Z","lastTransitionTime":"2026-02-02T15:16:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:43 crc kubenswrapper[4733]: I0202 15:16:43.190759 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:43 crc kubenswrapper[4733]: I0202 15:16:43.190829 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:43 crc kubenswrapper[4733]: I0202 15:16:43.190852 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:43 crc kubenswrapper[4733]: I0202 15:16:43.190882 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:43 crc kubenswrapper[4733]: I0202 15:16:43.190904 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:43Z","lastTransitionTime":"2026-02-02T15:16:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:43 crc kubenswrapper[4733]: I0202 15:16:43.250790 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-27 12:52:50.452872804 +0000 UTC Feb 02 15:16:43 crc kubenswrapper[4733]: I0202 15:16:43.254204 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:16:43 crc kubenswrapper[4733]: I0202 15:16:43.254322 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:16:43 crc kubenswrapper[4733]: I0202 15:16:43.254224 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:16:43 crc kubenswrapper[4733]: I0202 15:16:43.254328 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:16:43 crc kubenswrapper[4733]: E0202 15:16:43.254830 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:16:43 crc kubenswrapper[4733]: E0202 15:16:43.255021 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:16:43 crc kubenswrapper[4733]: E0202 15:16:43.255083 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:16:43 crc kubenswrapper[4733]: E0202 15:16:43.255292 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:16:43 crc kubenswrapper[4733]: I0202 15:16:43.293731 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:43 crc kubenswrapper[4733]: I0202 15:16:43.293774 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:43 crc kubenswrapper[4733]: I0202 15:16:43.293785 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:43 crc kubenswrapper[4733]: I0202 15:16:43.293804 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:43 crc kubenswrapper[4733]: I0202 15:16:43.293817 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:43Z","lastTransitionTime":"2026-02-02T15:16:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:43 crc kubenswrapper[4733]: I0202 15:16:43.397061 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:43 crc kubenswrapper[4733]: I0202 15:16:43.397133 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:43 crc kubenswrapper[4733]: I0202 15:16:43.397157 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:43 crc kubenswrapper[4733]: I0202 15:16:43.397237 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:43 crc kubenswrapper[4733]: I0202 15:16:43.397261 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:43Z","lastTransitionTime":"2026-02-02T15:16:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:43 crc kubenswrapper[4733]: I0202 15:16:43.500008 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:43 crc kubenswrapper[4733]: I0202 15:16:43.500071 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:43 crc kubenswrapper[4733]: I0202 15:16:43.500096 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:43 crc kubenswrapper[4733]: I0202 15:16:43.500127 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:43 crc kubenswrapper[4733]: I0202 15:16:43.500149 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:43Z","lastTransitionTime":"2026-02-02T15:16:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:43 crc kubenswrapper[4733]: I0202 15:16:43.603468 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:43 crc kubenswrapper[4733]: I0202 15:16:43.603562 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:43 crc kubenswrapper[4733]: I0202 15:16:43.603591 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:43 crc kubenswrapper[4733]: I0202 15:16:43.603622 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:43 crc kubenswrapper[4733]: I0202 15:16:43.603644 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:43Z","lastTransitionTime":"2026-02-02T15:16:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:43 crc kubenswrapper[4733]: I0202 15:16:43.706932 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:43 crc kubenswrapper[4733]: I0202 15:16:43.707017 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:43 crc kubenswrapper[4733]: I0202 15:16:43.707046 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:43 crc kubenswrapper[4733]: I0202 15:16:43.707080 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:43 crc kubenswrapper[4733]: I0202 15:16:43.707103 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:43Z","lastTransitionTime":"2026-02-02T15:16:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:43 crc kubenswrapper[4733]: I0202 15:16:43.809877 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:43 crc kubenswrapper[4733]: I0202 15:16:43.809954 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:43 crc kubenswrapper[4733]: I0202 15:16:43.809979 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:43 crc kubenswrapper[4733]: I0202 15:16:43.810012 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:43 crc kubenswrapper[4733]: I0202 15:16:43.810034 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:43Z","lastTransitionTime":"2026-02-02T15:16:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:43 crc kubenswrapper[4733]: I0202 15:16:43.912720 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:43 crc kubenswrapper[4733]: I0202 15:16:43.912759 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:43 crc kubenswrapper[4733]: I0202 15:16:43.912769 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:43 crc kubenswrapper[4733]: I0202 15:16:43.912785 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:43 crc kubenswrapper[4733]: I0202 15:16:43.912797 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:43Z","lastTransitionTime":"2026-02-02T15:16:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:44 crc kubenswrapper[4733]: I0202 15:16:44.015070 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:44 crc kubenswrapper[4733]: I0202 15:16:44.015127 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:44 crc kubenswrapper[4733]: I0202 15:16:44.015144 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:44 crc kubenswrapper[4733]: I0202 15:16:44.015197 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:44 crc kubenswrapper[4733]: I0202 15:16:44.015223 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:44Z","lastTransitionTime":"2026-02-02T15:16:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:44 crc kubenswrapper[4733]: I0202 15:16:44.117507 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:44 crc kubenswrapper[4733]: I0202 15:16:44.117760 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:44 crc kubenswrapper[4733]: I0202 15:16:44.117842 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:44 crc kubenswrapper[4733]: I0202 15:16:44.117921 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:44 crc kubenswrapper[4733]: I0202 15:16:44.117992 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:44Z","lastTransitionTime":"2026-02-02T15:16:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:44 crc kubenswrapper[4733]: I0202 15:16:44.221598 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:44 crc kubenswrapper[4733]: I0202 15:16:44.221674 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:44 crc kubenswrapper[4733]: I0202 15:16:44.221699 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:44 crc kubenswrapper[4733]: I0202 15:16:44.221728 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:44 crc kubenswrapper[4733]: I0202 15:16:44.221749 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:44Z","lastTransitionTime":"2026-02-02T15:16:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:44 crc kubenswrapper[4733]: I0202 15:16:44.251710 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-24 22:44:13.321703823 +0000 UTC Feb 02 15:16:44 crc kubenswrapper[4733]: I0202 15:16:44.325273 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:44 crc kubenswrapper[4733]: I0202 15:16:44.325352 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:44 crc kubenswrapper[4733]: I0202 15:16:44.325424 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:44 crc kubenswrapper[4733]: I0202 15:16:44.325455 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:44 crc kubenswrapper[4733]: I0202 15:16:44.325482 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:44Z","lastTransitionTime":"2026-02-02T15:16:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:44 crc kubenswrapper[4733]: I0202 15:16:44.428362 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:44 crc kubenswrapper[4733]: I0202 15:16:44.428418 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:44 crc kubenswrapper[4733]: I0202 15:16:44.428434 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:44 crc kubenswrapper[4733]: I0202 15:16:44.428456 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:44 crc kubenswrapper[4733]: I0202 15:16:44.428473 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:44Z","lastTransitionTime":"2026-02-02T15:16:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:44 crc kubenswrapper[4733]: I0202 15:16:44.531540 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:44 crc kubenswrapper[4733]: I0202 15:16:44.531584 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:44 crc kubenswrapper[4733]: I0202 15:16:44.531592 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:44 crc kubenswrapper[4733]: I0202 15:16:44.531608 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:44 crc kubenswrapper[4733]: I0202 15:16:44.531617 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:44Z","lastTransitionTime":"2026-02-02T15:16:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:44 crc kubenswrapper[4733]: I0202 15:16:44.634123 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:44 crc kubenswrapper[4733]: I0202 15:16:44.634227 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:44 crc kubenswrapper[4733]: I0202 15:16:44.634247 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:44 crc kubenswrapper[4733]: I0202 15:16:44.634271 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:44 crc kubenswrapper[4733]: I0202 15:16:44.634289 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:44Z","lastTransitionTime":"2026-02-02T15:16:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:44 crc kubenswrapper[4733]: I0202 15:16:44.737294 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:44 crc kubenswrapper[4733]: I0202 15:16:44.737352 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:44 crc kubenswrapper[4733]: I0202 15:16:44.737368 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:44 crc kubenswrapper[4733]: I0202 15:16:44.737390 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:44 crc kubenswrapper[4733]: I0202 15:16:44.737406 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:44Z","lastTransitionTime":"2026-02-02T15:16:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:44 crc kubenswrapper[4733]: I0202 15:16:44.840617 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:44 crc kubenswrapper[4733]: I0202 15:16:44.840682 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:44 crc kubenswrapper[4733]: I0202 15:16:44.840703 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:44 crc kubenswrapper[4733]: I0202 15:16:44.840732 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:44 crc kubenswrapper[4733]: I0202 15:16:44.840754 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:44Z","lastTransitionTime":"2026-02-02T15:16:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:44 crc kubenswrapper[4733]: I0202 15:16:44.943701 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:44 crc kubenswrapper[4733]: I0202 15:16:44.943781 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:44 crc kubenswrapper[4733]: I0202 15:16:44.943804 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:44 crc kubenswrapper[4733]: I0202 15:16:44.943831 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:44 crc kubenswrapper[4733]: I0202 15:16:44.943851 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:44Z","lastTransitionTime":"2026-02-02T15:16:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.047052 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.047111 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.047119 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.047133 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.047143 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:45Z","lastTransitionTime":"2026-02-02T15:16:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.149743 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.149801 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.149820 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.149845 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.149862 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:45Z","lastTransitionTime":"2026-02-02T15:16:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.251998 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-14 18:22:35.617412568 +0000 UTC Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.252649 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.252697 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.252714 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.252740 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.252757 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:45Z","lastTransitionTime":"2026-02-02T15:16:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.255241 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.255442 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.255525 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.255588 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:16:45 crc kubenswrapper[4733]: E0202 15:16:45.255658 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:16:45 crc kubenswrapper[4733]: E0202 15:16:45.255748 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:16:45 crc kubenswrapper[4733]: E0202 15:16:45.255986 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:16:45 crc kubenswrapper[4733]: E0202 15:16:45.256256 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.356477 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.356535 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.356553 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.356577 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.356593 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:45Z","lastTransitionTime":"2026-02-02T15:16:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.459779 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.460095 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.460115 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.460133 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.460144 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:45Z","lastTransitionTime":"2026-02-02T15:16:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.563187 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.563474 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.563486 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.563501 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.563510 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:45Z","lastTransitionTime":"2026-02-02T15:16:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.666452 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.666511 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.666528 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.666553 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.666571 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:45Z","lastTransitionTime":"2026-02-02T15:16:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.769557 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.769617 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.769634 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.769662 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.769679 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:45Z","lastTransitionTime":"2026-02-02T15:16:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.872390 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.872452 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.872470 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.872495 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.872517 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:45Z","lastTransitionTime":"2026-02-02T15:16:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.975204 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.975246 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.975256 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.975271 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:45 crc kubenswrapper[4733]: I0202 15:16:45.975280 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:45Z","lastTransitionTime":"2026-02-02T15:16:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:46 crc kubenswrapper[4733]: I0202 15:16:46.078230 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:46 crc kubenswrapper[4733]: I0202 15:16:46.078279 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:46 crc kubenswrapper[4733]: I0202 15:16:46.078291 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:46 crc kubenswrapper[4733]: I0202 15:16:46.078310 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:46 crc kubenswrapper[4733]: I0202 15:16:46.078324 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:46Z","lastTransitionTime":"2026-02-02T15:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:46 crc kubenswrapper[4733]: I0202 15:16:46.181570 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:46 crc kubenswrapper[4733]: I0202 15:16:46.181646 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:46 crc kubenswrapper[4733]: I0202 15:16:46.181665 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:46 crc kubenswrapper[4733]: I0202 15:16:46.181693 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:46 crc kubenswrapper[4733]: I0202 15:16:46.181714 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:46Z","lastTransitionTime":"2026-02-02T15:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:46 crc kubenswrapper[4733]: I0202 15:16:46.252384 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-27 15:48:58.987820286 +0000 UTC Feb 02 15:16:46 crc kubenswrapper[4733]: I0202 15:16:46.284347 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:46 crc kubenswrapper[4733]: I0202 15:16:46.284426 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:46 crc kubenswrapper[4733]: I0202 15:16:46.284451 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:46 crc kubenswrapper[4733]: I0202 15:16:46.284483 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:46 crc kubenswrapper[4733]: I0202 15:16:46.284506 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:46Z","lastTransitionTime":"2026-02-02T15:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:46 crc kubenswrapper[4733]: I0202 15:16:46.387351 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:46 crc kubenswrapper[4733]: I0202 15:16:46.387426 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:46 crc kubenswrapper[4733]: I0202 15:16:46.387445 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:46 crc kubenswrapper[4733]: I0202 15:16:46.387471 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:46 crc kubenswrapper[4733]: I0202 15:16:46.387489 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:46Z","lastTransitionTime":"2026-02-02T15:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:46 crc kubenswrapper[4733]: I0202 15:16:46.490052 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:46 crc kubenswrapper[4733]: I0202 15:16:46.490099 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:46 crc kubenswrapper[4733]: I0202 15:16:46.490109 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:46 crc kubenswrapper[4733]: I0202 15:16:46.490123 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:46 crc kubenswrapper[4733]: I0202 15:16:46.490134 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:46Z","lastTransitionTime":"2026-02-02T15:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:46 crc kubenswrapper[4733]: I0202 15:16:46.593471 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:46 crc kubenswrapper[4733]: I0202 15:16:46.593532 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:46 crc kubenswrapper[4733]: I0202 15:16:46.593548 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:46 crc kubenswrapper[4733]: I0202 15:16:46.593572 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:46 crc kubenswrapper[4733]: I0202 15:16:46.593593 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:46Z","lastTransitionTime":"2026-02-02T15:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:46 crc kubenswrapper[4733]: I0202 15:16:46.696792 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:46 crc kubenswrapper[4733]: I0202 15:16:46.696853 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:46 crc kubenswrapper[4733]: I0202 15:16:46.696872 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:46 crc kubenswrapper[4733]: I0202 15:16:46.696897 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:46 crc kubenswrapper[4733]: I0202 15:16:46.696916 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:46Z","lastTransitionTime":"2026-02-02T15:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:46 crc kubenswrapper[4733]: I0202 15:16:46.800329 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:46 crc kubenswrapper[4733]: I0202 15:16:46.800406 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:46 crc kubenswrapper[4733]: I0202 15:16:46.800425 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:46 crc kubenswrapper[4733]: I0202 15:16:46.800451 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:46 crc kubenswrapper[4733]: I0202 15:16:46.800469 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:46Z","lastTransitionTime":"2026-02-02T15:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:46 crc kubenswrapper[4733]: I0202 15:16:46.904149 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:46 crc kubenswrapper[4733]: I0202 15:16:46.904248 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:46 crc kubenswrapper[4733]: I0202 15:16:46.904266 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:46 crc kubenswrapper[4733]: I0202 15:16:46.904292 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:46 crc kubenswrapper[4733]: I0202 15:16:46.904312 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:46Z","lastTransitionTime":"2026-02-02T15:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.012743 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.012802 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.012821 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.012846 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.012865 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:47Z","lastTransitionTime":"2026-02-02T15:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.115963 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.116019 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.116037 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.116060 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.116078 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:47Z","lastTransitionTime":"2026-02-02T15:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.219078 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.219154 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.219217 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.219263 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.219287 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:47Z","lastTransitionTime":"2026-02-02T15:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.252892 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 17:58:14.306460352 +0000 UTC Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.254366 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.254383 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.254473 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:16:47 crc kubenswrapper[4733]: E0202 15:16:47.254517 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.254554 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:16:47 crc kubenswrapper[4733]: E0202 15:16:47.254683 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:16:47 crc kubenswrapper[4733]: E0202 15:16:47.254747 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:16:47 crc kubenswrapper[4733]: E0202 15:16:47.254806 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.278637 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podStartSLOduration=71.278603781 podStartE2EDuration="1m11.278603781s" podCreationTimestamp="2026-02-02 15:15:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:16:47.278602061 +0000 UTC m=+90.730063459" watchObservedRunningTime="2026-02-02 15:16:47.278603781 +0000 UTC m=+90.730065179" Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.322351 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.322413 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.322428 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.322445 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.322458 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:47Z","lastTransitionTime":"2026-02-02T15:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.347382 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-jh6vk" podStartSLOduration=71.347357415 podStartE2EDuration="1m11.347357415s" podCreationTimestamp="2026-02-02 15:15:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:16:47.328525441 +0000 UTC m=+90.779986839" watchObservedRunningTime="2026-02-02 15:16:47.347357415 +0000 UTC m=+90.798818783" Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.397588 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-5xgpn" podStartSLOduration=71.397551682 podStartE2EDuration="1m11.397551682s" podCreationTimestamp="2026-02-02 15:15:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:16:47.383438792 +0000 UTC m=+90.834900150" watchObservedRunningTime="2026-02-02 15:16:47.397551682 +0000 UTC m=+90.849013040" Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.424409 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.424458 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.424471 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.424491 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.424505 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:47Z","lastTransitionTime":"2026-02-02T15:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.428057 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=15.428042282 podStartE2EDuration="15.428042282s" podCreationTimestamp="2026-02-02 15:16:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:16:47.426903242 +0000 UTC m=+90.878364630" watchObservedRunningTime="2026-02-02 15:16:47.428042282 +0000 UTC m=+90.879503640" Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.428389 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=70.428382551 podStartE2EDuration="1m10.428382551s" podCreationTimestamp="2026-02-02 15:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:16:47.410709927 +0000 UTC m=+90.862171295" watchObservedRunningTime="2026-02-02 15:16:47.428382551 +0000 UTC m=+90.879843909" Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.451486 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-8c597" podStartSLOduration=70.451463177 podStartE2EDuration="1m10.451463177s" podCreationTimestamp="2026-02-02 15:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:16:47.450891122 +0000 UTC m=+90.902352480" watchObservedRunningTime="2026-02-02 15:16:47.451463177 +0000 UTC m=+90.902924535" Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.487762 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hbm26" podStartSLOduration=70.487739029 podStartE2EDuration="1m10.487739029s" podCreationTimestamp="2026-02-02 15:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:16:47.462891867 +0000 UTC m=+90.914353225" watchObservedRunningTime="2026-02-02 15:16:47.487739029 +0000 UTC m=+90.939200387" Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.488108 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=70.488102708 podStartE2EDuration="1m10.488102708s" podCreationTimestamp="2026-02-02 15:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:16:47.486940388 +0000 UTC m=+90.938401756" watchObservedRunningTime="2026-02-02 15:16:47.488102708 +0000 UTC m=+90.939564076" Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.505797 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-nr6ps" podStartSLOduration=71.505766902 podStartE2EDuration="1m11.505766902s" podCreationTimestamp="2026-02-02 15:15:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:16:47.504272863 +0000 UTC m=+90.955734221" watchObservedRunningTime="2026-02-02 15:16:47.505766902 +0000 UTC m=+90.957228310" Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.527578 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.527623 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.527635 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.527653 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.527665 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:47Z","lastTransitionTime":"2026-02-02T15:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.561574 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=68.561553816 podStartE2EDuration="1m8.561553816s" podCreationTimestamp="2026-02-02 15:15:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:16:47.546620204 +0000 UTC m=+90.998081562" watchObservedRunningTime="2026-02-02 15:16:47.561553816 +0000 UTC m=+91.013015164" Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.630933 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.631014 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.631034 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.631061 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.631080 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:47Z","lastTransitionTime":"2026-02-02T15:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.733918 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.733981 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.734003 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.734031 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.734056 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:47Z","lastTransitionTime":"2026-02-02T15:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.836711 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.836760 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.836776 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.836798 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.836816 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:47Z","lastTransitionTime":"2026-02-02T15:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.939244 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.939288 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.939308 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.939327 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:47 crc kubenswrapper[4733]: I0202 15:16:47.939340 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:47Z","lastTransitionTime":"2026-02-02T15:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:48 crc kubenswrapper[4733]: I0202 15:16:48.042146 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:48 crc kubenswrapper[4733]: I0202 15:16:48.042279 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:48 crc kubenswrapper[4733]: I0202 15:16:48.042299 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:48 crc kubenswrapper[4733]: I0202 15:16:48.042323 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:48 crc kubenswrapper[4733]: I0202 15:16:48.042340 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:48Z","lastTransitionTime":"2026-02-02T15:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:48 crc kubenswrapper[4733]: I0202 15:16:48.145113 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:48 crc kubenswrapper[4733]: I0202 15:16:48.145196 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:48 crc kubenswrapper[4733]: I0202 15:16:48.145214 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:48 crc kubenswrapper[4733]: I0202 15:16:48.145236 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:48 crc kubenswrapper[4733]: I0202 15:16:48.145254 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:48Z","lastTransitionTime":"2026-02-02T15:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:48 crc kubenswrapper[4733]: I0202 15:16:48.248010 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:48 crc kubenswrapper[4733]: I0202 15:16:48.248090 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:48 crc kubenswrapper[4733]: I0202 15:16:48.248109 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:48 crc kubenswrapper[4733]: I0202 15:16:48.248130 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:48 crc kubenswrapper[4733]: I0202 15:16:48.248145 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:48Z","lastTransitionTime":"2026-02-02T15:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:48 crc kubenswrapper[4733]: I0202 15:16:48.253235 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 23:56:44.59859895 +0000 UTC Feb 02 15:16:48 crc kubenswrapper[4733]: I0202 15:16:48.284022 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 02 15:16:48 crc kubenswrapper[4733]: I0202 15:16:48.284064 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 02 15:16:48 crc kubenswrapper[4733]: I0202 15:16:48.284079 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 02 15:16:48 crc kubenswrapper[4733]: I0202 15:16:48.284098 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 02 15:16:48 crc kubenswrapper[4733]: I0202 15:16:48.284113 4733 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-02T15:16:48Z","lastTransitionTime":"2026-02-02T15:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 02 15:16:48 crc kubenswrapper[4733]: I0202 15:16:48.343982 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=35.343949407 podStartE2EDuration="35.343949407s" podCreationTimestamp="2026-02-02 15:16:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:16:47.562755117 +0000 UTC m=+91.014216475" watchObservedRunningTime="2026-02-02 15:16:48.343949407 +0000 UTC m=+91.795410805" Feb 02 15:16:48 crc kubenswrapper[4733]: I0202 15:16:48.345508 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-x56pt"] Feb 02 15:16:48 crc kubenswrapper[4733]: I0202 15:16:48.351365 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x56pt" Feb 02 15:16:48 crc kubenswrapper[4733]: I0202 15:16:48.354037 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Feb 02 15:16:48 crc kubenswrapper[4733]: I0202 15:16:48.354411 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Feb 02 15:16:48 crc kubenswrapper[4733]: I0202 15:16:48.354625 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Feb 02 15:16:48 crc kubenswrapper[4733]: I0202 15:16:48.356528 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Feb 02 15:16:48 crc kubenswrapper[4733]: I0202 15:16:48.501950 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/94749557-42ae-4928-9d1e-d6d259f012a3-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-x56pt\" (UID: \"94749557-42ae-4928-9d1e-d6d259f012a3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x56pt" Feb 02 15:16:48 crc kubenswrapper[4733]: I0202 15:16:48.502007 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/94749557-42ae-4928-9d1e-d6d259f012a3-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-x56pt\" (UID: \"94749557-42ae-4928-9d1e-d6d259f012a3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x56pt" Feb 02 15:16:48 crc kubenswrapper[4733]: I0202 15:16:48.502069 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/94749557-42ae-4928-9d1e-d6d259f012a3-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-x56pt\" (UID: \"94749557-42ae-4928-9d1e-d6d259f012a3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x56pt" Feb 02 15:16:48 crc kubenswrapper[4733]: I0202 15:16:48.502197 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/94749557-42ae-4928-9d1e-d6d259f012a3-service-ca\") pod \"cluster-version-operator-5c965bbfc6-x56pt\" (UID: \"94749557-42ae-4928-9d1e-d6d259f012a3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x56pt" Feb 02 15:16:48 crc kubenswrapper[4733]: I0202 15:16:48.502274 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/94749557-42ae-4928-9d1e-d6d259f012a3-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-x56pt\" (UID: \"94749557-42ae-4928-9d1e-d6d259f012a3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x56pt" Feb 02 15:16:48 crc kubenswrapper[4733]: I0202 15:16:48.603455 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/94749557-42ae-4928-9d1e-d6d259f012a3-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-x56pt\" (UID: \"94749557-42ae-4928-9d1e-d6d259f012a3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x56pt" Feb 02 15:16:48 crc kubenswrapper[4733]: I0202 15:16:48.603597 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/94749557-42ae-4928-9d1e-d6d259f012a3-service-ca\") pod \"cluster-version-operator-5c965bbfc6-x56pt\" (UID: \"94749557-42ae-4928-9d1e-d6d259f012a3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x56pt" Feb 02 15:16:48 crc kubenswrapper[4733]: I0202 15:16:48.603655 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/94749557-42ae-4928-9d1e-d6d259f012a3-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-x56pt\" (UID: \"94749557-42ae-4928-9d1e-d6d259f012a3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x56pt" Feb 02 15:16:48 crc kubenswrapper[4733]: I0202 15:16:48.604011 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/94749557-42ae-4928-9d1e-d6d259f012a3-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-x56pt\" (UID: \"94749557-42ae-4928-9d1e-d6d259f012a3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x56pt" Feb 02 15:16:48 crc kubenswrapper[4733]: I0202 15:16:48.604330 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/94749557-42ae-4928-9d1e-d6d259f012a3-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-x56pt\" (UID: \"94749557-42ae-4928-9d1e-d6d259f012a3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x56pt" Feb 02 15:16:48 crc kubenswrapper[4733]: I0202 15:16:48.604487 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/94749557-42ae-4928-9d1e-d6d259f012a3-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-x56pt\" (UID: \"94749557-42ae-4928-9d1e-d6d259f012a3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x56pt" Feb 02 15:16:48 crc kubenswrapper[4733]: I0202 15:16:48.604658 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/94749557-42ae-4928-9d1e-d6d259f012a3-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-x56pt\" (UID: \"94749557-42ae-4928-9d1e-d6d259f012a3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x56pt" Feb 02 15:16:48 crc kubenswrapper[4733]: I0202 15:16:48.605416 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/94749557-42ae-4928-9d1e-d6d259f012a3-service-ca\") pod \"cluster-version-operator-5c965bbfc6-x56pt\" (UID: \"94749557-42ae-4928-9d1e-d6d259f012a3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x56pt" Feb 02 15:16:48 crc kubenswrapper[4733]: I0202 15:16:48.614445 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/94749557-42ae-4928-9d1e-d6d259f012a3-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-x56pt\" (UID: \"94749557-42ae-4928-9d1e-d6d259f012a3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x56pt" Feb 02 15:16:48 crc kubenswrapper[4733]: I0202 15:16:48.634233 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/94749557-42ae-4928-9d1e-d6d259f012a3-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-x56pt\" (UID: \"94749557-42ae-4928-9d1e-d6d259f012a3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x56pt" Feb 02 15:16:48 crc kubenswrapper[4733]: I0202 15:16:48.674700 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x56pt" Feb 02 15:16:48 crc kubenswrapper[4733]: W0202 15:16:48.700448 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod94749557_42ae_4928_9d1e_d6d259f012a3.slice/crio-14a148baf76c1370f320ea2bc3cdc149481e0744a39c17d4729bc2fa528e178a WatchSource:0}: Error finding container 14a148baf76c1370f320ea2bc3cdc149481e0744a39c17d4729bc2fa528e178a: Status 404 returned error can't find the container with id 14a148baf76c1370f320ea2bc3cdc149481e0744a39c17d4729bc2fa528e178a Feb 02 15:16:48 crc kubenswrapper[4733]: I0202 15:16:48.816398 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x56pt" event={"ID":"94749557-42ae-4928-9d1e-d6d259f012a3","Type":"ContainerStarted","Data":"14a148baf76c1370f320ea2bc3cdc149481e0744a39c17d4729bc2fa528e178a"} Feb 02 15:16:49 crc kubenswrapper[4733]: I0202 15:16:49.254388 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-04 12:39:48.19528657 +0000 UTC Feb 02 15:16:49 crc kubenswrapper[4733]: I0202 15:16:49.254480 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Feb 02 15:16:49 crc kubenswrapper[4733]: I0202 15:16:49.254662 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:16:49 crc kubenswrapper[4733]: I0202 15:16:49.254685 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:16:49 crc kubenswrapper[4733]: I0202 15:16:49.254770 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:16:49 crc kubenswrapper[4733]: I0202 15:16:49.254800 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:16:49 crc kubenswrapper[4733]: E0202 15:16:49.254944 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:16:49 crc kubenswrapper[4733]: E0202 15:16:49.255060 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:16:49 crc kubenswrapper[4733]: E0202 15:16:49.255236 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:16:49 crc kubenswrapper[4733]: E0202 15:16:49.255478 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:16:49 crc kubenswrapper[4733]: I0202 15:16:49.269345 4733 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Feb 02 15:16:49 crc kubenswrapper[4733]: I0202 15:16:49.821926 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x56pt" event={"ID":"94749557-42ae-4928-9d1e-d6d259f012a3","Type":"ContainerStarted","Data":"4b3edc3895c090ce3f671f68c8b91e16c3755b63de3affeb8c7f249b4471771d"} Feb 02 15:16:49 crc kubenswrapper[4733]: I0202 15:16:49.840466 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x56pt" podStartSLOduration=72.840441775 podStartE2EDuration="1m12.840441775s" podCreationTimestamp="2026-02-02 15:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:16:49.839792228 +0000 UTC m=+93.291253636" watchObservedRunningTime="2026-02-02 15:16:49.840441775 +0000 UTC m=+93.291903163" Feb 02 15:16:51 crc kubenswrapper[4733]: I0202 15:16:51.254818 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:16:51 crc kubenswrapper[4733]: I0202 15:16:51.254933 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:16:51 crc kubenswrapper[4733]: E0202 15:16:51.254970 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:16:51 crc kubenswrapper[4733]: I0202 15:16:51.255090 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:16:51 crc kubenswrapper[4733]: E0202 15:16:51.255249 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:16:51 crc kubenswrapper[4733]: I0202 15:16:51.255311 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:16:51 crc kubenswrapper[4733]: E0202 15:16:51.255395 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:16:51 crc kubenswrapper[4733]: E0202 15:16:51.255487 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:16:53 crc kubenswrapper[4733]: I0202 15:16:53.254502 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:16:53 crc kubenswrapper[4733]: I0202 15:16:53.254627 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:16:53 crc kubenswrapper[4733]: E0202 15:16:53.254685 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:16:53 crc kubenswrapper[4733]: I0202 15:16:53.254716 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:16:53 crc kubenswrapper[4733]: I0202 15:16:53.254834 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:16:53 crc kubenswrapper[4733]: E0202 15:16:53.255019 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:16:53 crc kubenswrapper[4733]: E0202 15:16:53.255134 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:16:53 crc kubenswrapper[4733]: E0202 15:16:53.255274 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:16:55 crc kubenswrapper[4733]: I0202 15:16:55.254351 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:16:55 crc kubenswrapper[4733]: I0202 15:16:55.254472 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:16:55 crc kubenswrapper[4733]: I0202 15:16:55.254532 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:16:55 crc kubenswrapper[4733]: I0202 15:16:55.254550 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:16:55 crc kubenswrapper[4733]: E0202 15:16:55.255542 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:16:55 crc kubenswrapper[4733]: E0202 15:16:55.255703 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:16:55 crc kubenswrapper[4733]: E0202 15:16:55.255878 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:16:55 crc kubenswrapper[4733]: E0202 15:16:55.256010 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:16:56 crc kubenswrapper[4733]: I0202 15:16:56.255404 4733 scope.go:117] "RemoveContainer" containerID="55b2c8f5f7c8b659a1b54bb592229f81596804f972fc4658e1f867054721257d" Feb 02 15:16:56 crc kubenswrapper[4733]: E0202 15:16:56.255560 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-qxvkf_openshift-ovn-kubernetes(f5a81ada-e393-499a-aac6-7b40349958bb)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" Feb 02 15:16:57 crc kubenswrapper[4733]: I0202 15:16:57.256300 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:16:57 crc kubenswrapper[4733]: E0202 15:16:57.256720 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:16:57 crc kubenswrapper[4733]: I0202 15:16:57.256967 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:16:57 crc kubenswrapper[4733]: E0202 15:16:57.257050 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:16:57 crc kubenswrapper[4733]: I0202 15:16:57.257278 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:16:57 crc kubenswrapper[4733]: E0202 15:16:57.257391 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:16:57 crc kubenswrapper[4733]: I0202 15:16:57.257585 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:16:57 crc kubenswrapper[4733]: E0202 15:16:57.257679 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:16:59 crc kubenswrapper[4733]: I0202 15:16:59.255241 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:16:59 crc kubenswrapper[4733]: I0202 15:16:59.255568 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:16:59 crc kubenswrapper[4733]: I0202 15:16:59.255626 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:16:59 crc kubenswrapper[4733]: I0202 15:16:59.255989 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:16:59 crc kubenswrapper[4733]: E0202 15:16:59.256118 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:16:59 crc kubenswrapper[4733]: E0202 15:16:59.256245 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:16:59 crc kubenswrapper[4733]: E0202 15:16:59.256327 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:16:59 crc kubenswrapper[4733]: E0202 15:16:59.256439 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:17:01 crc kubenswrapper[4733]: I0202 15:17:01.254587 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:17:01 crc kubenswrapper[4733]: I0202 15:17:01.254673 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:17:01 crc kubenswrapper[4733]: I0202 15:17:01.254587 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:17:01 crc kubenswrapper[4733]: E0202 15:17:01.254815 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:17:01 crc kubenswrapper[4733]: E0202 15:17:01.255010 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:17:01 crc kubenswrapper[4733]: E0202 15:17:01.255135 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:17:01 crc kubenswrapper[4733]: I0202 15:17:01.255377 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:17:01 crc kubenswrapper[4733]: E0202 15:17:01.255547 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:17:03 crc kubenswrapper[4733]: I0202 15:17:03.255111 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:17:03 crc kubenswrapper[4733]: I0202 15:17:03.255120 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:17:03 crc kubenswrapper[4733]: E0202 15:17:03.255308 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:17:03 crc kubenswrapper[4733]: I0202 15:17:03.255220 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:17:03 crc kubenswrapper[4733]: I0202 15:17:03.255083 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:17:03 crc kubenswrapper[4733]: E0202 15:17:03.255440 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:17:03 crc kubenswrapper[4733]: E0202 15:17:03.255527 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:17:03 crc kubenswrapper[4733]: E0202 15:17:03.255616 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:17:05 crc kubenswrapper[4733]: I0202 15:17:05.254084 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:17:05 crc kubenswrapper[4733]: I0202 15:17:05.254134 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:17:05 crc kubenswrapper[4733]: I0202 15:17:05.254182 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:17:05 crc kubenswrapper[4733]: E0202 15:17:05.254326 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:17:05 crc kubenswrapper[4733]: I0202 15:17:05.254380 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:17:05 crc kubenswrapper[4733]: E0202 15:17:05.254587 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:17:05 crc kubenswrapper[4733]: E0202 15:17:05.254695 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:17:05 crc kubenswrapper[4733]: E0202 15:17:05.254870 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:17:07 crc kubenswrapper[4733]: I0202 15:17:07.254650 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:17:07 crc kubenswrapper[4733]: I0202 15:17:07.254800 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:17:07 crc kubenswrapper[4733]: E0202 15:17:07.256550 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:17:07 crc kubenswrapper[4733]: I0202 15:17:07.256636 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:17:07 crc kubenswrapper[4733]: I0202 15:17:07.256670 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:17:07 crc kubenswrapper[4733]: E0202 15:17:07.256858 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:17:07 crc kubenswrapper[4733]: E0202 15:17:07.256939 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:17:07 crc kubenswrapper[4733]: E0202 15:17:07.257104 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:17:07 crc kubenswrapper[4733]: I0202 15:17:07.258257 4733 scope.go:117] "RemoveContainer" containerID="55b2c8f5f7c8b659a1b54bb592229f81596804f972fc4658e1f867054721257d" Feb 02 15:17:07 crc kubenswrapper[4733]: E0202 15:17:07.258505 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-qxvkf_openshift-ovn-kubernetes(f5a81ada-e393-499a-aac6-7b40349958bb)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" Feb 02 15:17:09 crc kubenswrapper[4733]: I0202 15:17:09.254447 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:17:09 crc kubenswrapper[4733]: E0202 15:17:09.254630 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:17:09 crc kubenswrapper[4733]: I0202 15:17:09.254674 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:17:09 crc kubenswrapper[4733]: I0202 15:17:09.254740 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:17:09 crc kubenswrapper[4733]: I0202 15:17:09.254689 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:17:09 crc kubenswrapper[4733]: E0202 15:17:09.254885 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:17:09 crc kubenswrapper[4733]: E0202 15:17:09.255352 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:17:09 crc kubenswrapper[4733]: E0202 15:17:09.255794 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:17:11 crc kubenswrapper[4733]: I0202 15:17:11.254456 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:17:11 crc kubenswrapper[4733]: I0202 15:17:11.254558 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:17:11 crc kubenswrapper[4733]: E0202 15:17:11.254718 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:17:11 crc kubenswrapper[4733]: E0202 15:17:11.254819 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:17:11 crc kubenswrapper[4733]: I0202 15:17:11.254886 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:17:11 crc kubenswrapper[4733]: I0202 15:17:11.255466 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:17:11 crc kubenswrapper[4733]: E0202 15:17:11.255635 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:17:11 crc kubenswrapper[4733]: E0202 15:17:11.255788 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:17:11 crc kubenswrapper[4733]: I0202 15:17:11.903700 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jh6vk_b76a05d6-d9c7-4183-ad90-b3131f225841/kube-multus/1.log" Feb 02 15:17:11 crc kubenswrapper[4733]: I0202 15:17:11.904668 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jh6vk_b76a05d6-d9c7-4183-ad90-b3131f225841/kube-multus/0.log" Feb 02 15:17:11 crc kubenswrapper[4733]: I0202 15:17:11.904750 4733 generic.go:334] "Generic (PLEG): container finished" podID="b76a05d6-d9c7-4183-ad90-b3131f225841" containerID="f57aea989227fdbd7bb59c51fade50d1076929f6288d060808b51bd1bdd7df04" exitCode=1 Feb 02 15:17:11 crc kubenswrapper[4733]: I0202 15:17:11.904814 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jh6vk" event={"ID":"b76a05d6-d9c7-4183-ad90-b3131f225841","Type":"ContainerDied","Data":"f57aea989227fdbd7bb59c51fade50d1076929f6288d060808b51bd1bdd7df04"} Feb 02 15:17:11 crc kubenswrapper[4733]: I0202 15:17:11.904897 4733 scope.go:117] "RemoveContainer" containerID="7f2659e2a914d3191daddbfb5ad109e487f03dabb559d324244e127add7d41d9" Feb 02 15:17:11 crc kubenswrapper[4733]: I0202 15:17:11.905635 4733 scope.go:117] "RemoveContainer" containerID="f57aea989227fdbd7bb59c51fade50d1076929f6288d060808b51bd1bdd7df04" Feb 02 15:17:11 crc kubenswrapper[4733]: E0202 15:17:11.905963 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-jh6vk_openshift-multus(b76a05d6-d9c7-4183-ad90-b3131f225841)\"" pod="openshift-multus/multus-jh6vk" podUID="b76a05d6-d9c7-4183-ad90-b3131f225841" Feb 02 15:17:12 crc kubenswrapper[4733]: I0202 15:17:12.910715 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jh6vk_b76a05d6-d9c7-4183-ad90-b3131f225841/kube-multus/1.log" Feb 02 15:17:13 crc kubenswrapper[4733]: I0202 15:17:13.255022 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:17:13 crc kubenswrapper[4733]: I0202 15:17:13.255095 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:17:13 crc kubenswrapper[4733]: E0202 15:17:13.255215 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:17:13 crc kubenswrapper[4733]: I0202 15:17:13.255235 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:17:13 crc kubenswrapper[4733]: I0202 15:17:13.255263 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:17:13 crc kubenswrapper[4733]: E0202 15:17:13.255382 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:17:13 crc kubenswrapper[4733]: E0202 15:17:13.255577 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:17:13 crc kubenswrapper[4733]: E0202 15:17:13.255712 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:17:15 crc kubenswrapper[4733]: I0202 15:17:15.254193 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:17:15 crc kubenswrapper[4733]: I0202 15:17:15.254193 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:17:15 crc kubenswrapper[4733]: I0202 15:17:15.254293 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:17:15 crc kubenswrapper[4733]: I0202 15:17:15.254343 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:17:15 crc kubenswrapper[4733]: E0202 15:17:15.254932 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:17:15 crc kubenswrapper[4733]: E0202 15:17:15.255207 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:17:15 crc kubenswrapper[4733]: E0202 15:17:15.255384 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:17:15 crc kubenswrapper[4733]: E0202 15:17:15.255521 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:17:17 crc kubenswrapper[4733]: I0202 15:17:17.253975 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:17:17 crc kubenswrapper[4733]: I0202 15:17:17.254006 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:17:17 crc kubenswrapper[4733]: I0202 15:17:17.254054 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:17:17 crc kubenswrapper[4733]: E0202 15:17:17.256377 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:17:17 crc kubenswrapper[4733]: I0202 15:17:17.256410 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:17:17 crc kubenswrapper[4733]: E0202 15:17:17.257317 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:17:17 crc kubenswrapper[4733]: E0202 15:17:17.257359 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:17:17 crc kubenswrapper[4733]: E0202 15:17:17.257595 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:17:17 crc kubenswrapper[4733]: E0202 15:17:17.270159 4733 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Feb 02 15:17:17 crc kubenswrapper[4733]: E0202 15:17:17.355303 4733 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 02 15:17:19 crc kubenswrapper[4733]: I0202 15:17:19.254751 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:17:19 crc kubenswrapper[4733]: I0202 15:17:19.254854 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:17:19 crc kubenswrapper[4733]: I0202 15:17:19.254926 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:17:19 crc kubenswrapper[4733]: E0202 15:17:19.255014 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:17:19 crc kubenswrapper[4733]: I0202 15:17:19.255139 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:17:19 crc kubenswrapper[4733]: E0202 15:17:19.255234 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:17:19 crc kubenswrapper[4733]: E0202 15:17:19.255366 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:17:19 crc kubenswrapper[4733]: E0202 15:17:19.255508 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:17:19 crc kubenswrapper[4733]: I0202 15:17:19.256907 4733 scope.go:117] "RemoveContainer" containerID="55b2c8f5f7c8b659a1b54bb592229f81596804f972fc4658e1f867054721257d" Feb 02 15:17:19 crc kubenswrapper[4733]: I0202 15:17:19.938056 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxvkf_f5a81ada-e393-499a-aac6-7b40349958bb/ovnkube-controller/3.log" Feb 02 15:17:19 crc kubenswrapper[4733]: I0202 15:17:19.942849 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" event={"ID":"f5a81ada-e393-499a-aac6-7b40349958bb","Type":"ContainerStarted","Data":"21363125c857d52144029dc46a3a1553d0b588546a6aaeab91155f3802857fc5"} Feb 02 15:17:19 crc kubenswrapper[4733]: I0202 15:17:19.943447 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:17:19 crc kubenswrapper[4733]: I0202 15:17:19.990072 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" podStartSLOduration=103.990050778 podStartE2EDuration="1m43.990050778s" podCreationTimestamp="2026-02-02 15:15:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:17:19.98934983 +0000 UTC m=+123.440811228" watchObservedRunningTime="2026-02-02 15:17:19.990050778 +0000 UTC m=+123.441512146" Feb 02 15:17:20 crc kubenswrapper[4733]: I0202 15:17:20.219556 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-7r9fd"] Feb 02 15:17:20 crc kubenswrapper[4733]: I0202 15:17:20.219674 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:17:20 crc kubenswrapper[4733]: E0202 15:17:20.219758 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:17:21 crc kubenswrapper[4733]: I0202 15:17:21.254936 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:17:21 crc kubenswrapper[4733]: I0202 15:17:21.255002 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:17:21 crc kubenswrapper[4733]: E0202 15:17:21.255545 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:17:21 crc kubenswrapper[4733]: E0202 15:17:21.255611 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:17:21 crc kubenswrapper[4733]: I0202 15:17:21.255197 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:17:21 crc kubenswrapper[4733]: E0202 15:17:21.255722 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:17:22 crc kubenswrapper[4733]: I0202 15:17:22.255019 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:17:22 crc kubenswrapper[4733]: E0202 15:17:22.255258 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:17:22 crc kubenswrapper[4733]: E0202 15:17:22.356696 4733 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 02 15:17:23 crc kubenswrapper[4733]: I0202 15:17:23.254361 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:17:23 crc kubenswrapper[4733]: E0202 15:17:23.254537 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:17:23 crc kubenswrapper[4733]: I0202 15:17:23.254783 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:17:23 crc kubenswrapper[4733]: E0202 15:17:23.254989 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:17:23 crc kubenswrapper[4733]: I0202 15:17:23.255017 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:17:23 crc kubenswrapper[4733]: E0202 15:17:23.255377 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:17:24 crc kubenswrapper[4733]: I0202 15:17:24.254452 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:17:24 crc kubenswrapper[4733]: E0202 15:17:24.254648 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:17:25 crc kubenswrapper[4733]: I0202 15:17:25.254324 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:17:25 crc kubenswrapper[4733]: I0202 15:17:25.254359 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:17:25 crc kubenswrapper[4733]: E0202 15:17:25.254535 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:17:25 crc kubenswrapper[4733]: I0202 15:17:25.254595 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:17:25 crc kubenswrapper[4733]: E0202 15:17:25.254747 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:17:25 crc kubenswrapper[4733]: E0202 15:17:25.254881 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:17:26 crc kubenswrapper[4733]: I0202 15:17:26.254978 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:17:26 crc kubenswrapper[4733]: E0202 15:17:26.255196 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:17:26 crc kubenswrapper[4733]: I0202 15:17:26.255472 4733 scope.go:117] "RemoveContainer" containerID="f57aea989227fdbd7bb59c51fade50d1076929f6288d060808b51bd1bdd7df04" Feb 02 15:17:26 crc kubenswrapper[4733]: I0202 15:17:26.979151 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jh6vk_b76a05d6-d9c7-4183-ad90-b3131f225841/kube-multus/1.log" Feb 02 15:17:26 crc kubenswrapper[4733]: I0202 15:17:26.979621 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jh6vk" event={"ID":"b76a05d6-d9c7-4183-ad90-b3131f225841","Type":"ContainerStarted","Data":"0746ffd4e82c3554d4f63c139295a9cedb77073db5dda7ba15e32d519c0ed868"} Feb 02 15:17:27 crc kubenswrapper[4733]: I0202 15:17:27.254071 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:17:27 crc kubenswrapper[4733]: I0202 15:17:27.254077 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:17:27 crc kubenswrapper[4733]: I0202 15:17:27.254021 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:17:27 crc kubenswrapper[4733]: E0202 15:17:27.256039 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:17:27 crc kubenswrapper[4733]: E0202 15:17:27.256215 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:17:27 crc kubenswrapper[4733]: E0202 15:17:27.256409 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:17:27 crc kubenswrapper[4733]: E0202 15:17:27.357611 4733 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 02 15:17:28 crc kubenswrapper[4733]: I0202 15:17:28.254333 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:17:28 crc kubenswrapper[4733]: E0202 15:17:28.254854 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:17:29 crc kubenswrapper[4733]: I0202 15:17:29.254490 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:17:29 crc kubenswrapper[4733]: I0202 15:17:29.254537 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:17:29 crc kubenswrapper[4733]: E0202 15:17:29.254663 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:17:29 crc kubenswrapper[4733]: I0202 15:17:29.254748 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:17:29 crc kubenswrapper[4733]: E0202 15:17:29.254881 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:17:29 crc kubenswrapper[4733]: E0202 15:17:29.254981 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:17:30 crc kubenswrapper[4733]: I0202 15:17:30.254110 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:17:30 crc kubenswrapper[4733]: E0202 15:17:30.254408 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:17:31 crc kubenswrapper[4733]: I0202 15:17:31.254226 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:17:31 crc kubenswrapper[4733]: I0202 15:17:31.254271 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:17:31 crc kubenswrapper[4733]: I0202 15:17:31.254271 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:17:31 crc kubenswrapper[4733]: E0202 15:17:31.255695 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 02 15:17:31 crc kubenswrapper[4733]: E0202 15:17:31.255790 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 02 15:17:31 crc kubenswrapper[4733]: E0202 15:17:31.255876 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 02 15:17:32 crc kubenswrapper[4733]: I0202 15:17:32.254860 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:17:32 crc kubenswrapper[4733]: E0202 15:17:32.255074 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7r9fd" podUID="d4dfd44b-b363-4162-9f43-9a8dc4a3b110" Feb 02 15:17:33 crc kubenswrapper[4733]: I0202 15:17:33.254530 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:17:33 crc kubenswrapper[4733]: I0202 15:17:33.254591 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:17:33 crc kubenswrapper[4733]: I0202 15:17:33.254838 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:17:33 crc kubenswrapper[4733]: I0202 15:17:33.258928 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Feb 02 15:17:33 crc kubenswrapper[4733]: I0202 15:17:33.259069 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Feb 02 15:17:33 crc kubenswrapper[4733]: I0202 15:17:33.259100 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Feb 02 15:17:33 crc kubenswrapper[4733]: I0202 15:17:33.259199 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Feb 02 15:17:34 crc kubenswrapper[4733]: I0202 15:17:34.254031 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:17:34 crc kubenswrapper[4733]: I0202 15:17:34.266826 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Feb 02 15:17:34 crc kubenswrapper[4733]: I0202 15:17:34.266891 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.425826 4733 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.475321 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-79hfl"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.476485 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-79hfl" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.477261 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-74k4s"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.478099 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-74k4s" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.480800 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-mt8x9"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.481314 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-mt8x9" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.483870 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.487085 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.487412 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r8cnf"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.487483 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.487759 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.487971 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.488128 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r8cnf" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.488204 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.488434 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.488640 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.488812 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.488866 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-4cq4j"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.488989 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.489193 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.489475 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-4cq4j" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.491721 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-ct5l6"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.492297 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-ct5l6" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.494855 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-wd9qj"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.495519 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-m62b2"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.495855 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-m62b2" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.496271 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-wd9qj" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.498029 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.506081 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.507046 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.507541 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-98h6l"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.507750 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.508389 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-b68qv"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.508420 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-98h6l" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.509043 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-b68qv" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.510556 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/a3e78085-1f5d-4f0c-84bb-62b787b1fbdd-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-74k4s\" (UID: \"a3e78085-1f5d-4f0c-84bb-62b787b1fbdd\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-74k4s" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.510597 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dab28707-1fc8-45a4-8ad4-0842b7f70239-trusted-ca\") pod \"console-operator-58897d9998-ct5l6\" (UID: \"dab28707-1fc8-45a4-8ad4-0842b7f70239\") " pod="openshift-console-operator/console-operator-58897d9998-ct5l6" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.510622 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/d8fddc32-a0a6-4149-aa39-b7f51f78a772-machine-approver-tls\") pod \"machine-approver-56656f9798-79hfl\" (UID: \"d8fddc32-a0a6-4149-aa39-b7f51f78a772\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-79hfl" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.510646 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-4cq4j\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " pod="openshift-authentication/oauth-openshift-558db77b4-4cq4j" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.510671 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/b9dce599-89b8-4c6c-a12c-65f9ef64de9b-encryption-config\") pod \"apiserver-76f77b778f-wd9qj\" (UID: \"b9dce599-89b8-4c6c-a12c-65f9ef64de9b\") " pod="openshift-apiserver/apiserver-76f77b778f-wd9qj" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.510691 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzwg2\" (UniqueName: \"kubernetes.io/projected/b9dce599-89b8-4c6c-a12c-65f9ef64de9b-kube-api-access-lzwg2\") pod \"apiserver-76f77b778f-wd9qj\" (UID: \"b9dce599-89b8-4c6c-a12c-65f9ef64de9b\") " pod="openshift-apiserver/apiserver-76f77b778f-wd9qj" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.510710 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dab28707-1fc8-45a4-8ad4-0842b7f70239-config\") pod \"console-operator-58897d9998-ct5l6\" (UID: \"dab28707-1fc8-45a4-8ad4-0842b7f70239\") " pod="openshift-console-operator/console-operator-58897d9998-ct5l6" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.510741 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-f8l78"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.510743 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-4cq4j\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " pod="openshift-authentication/oauth-openshift-558db77b4-4cq4j" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.510934 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-audit-policies\") pod \"oauth-openshift-558db77b4-4cq4j\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " pod="openshift-authentication/oauth-openshift-558db77b4-4cq4j" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.510958 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b9dce599-89b8-4c6c-a12c-65f9ef64de9b-serving-cert\") pod \"apiserver-76f77b778f-wd9qj\" (UID: \"b9dce599-89b8-4c6c-a12c-65f9ef64de9b\") " pod="openshift-apiserver/apiserver-76f77b778f-wd9qj" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.510981 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/a3e78085-1f5d-4f0c-84bb-62b787b1fbdd-images\") pod \"machine-api-operator-5694c8668f-74k4s\" (UID: \"a3e78085-1f5d-4f0c-84bb-62b787b1fbdd\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-74k4s" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.511004 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-audit-dir\") pod \"oauth-openshift-558db77b4-4cq4j\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " pod="openshift-authentication/oauth-openshift-558db77b4-4cq4j" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.511035 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8c5w\" (UniqueName: \"kubernetes.io/projected/a16e3671-5260-4e88-8fcc-8ba2252af6ea-kube-api-access-g8c5w\") pod \"authentication-operator-69f744f599-mt8x9\" (UID: \"a16e3671-5260-4e88-8fcc-8ba2252af6ea\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mt8x9" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.511057 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-4cq4j\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " pod="openshift-authentication/oauth-openshift-558db77b4-4cq4j" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.511075 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7718b7b0-f1a7-4b87-8ebe-345e9100463d-config\") pod \"controller-manager-879f6c89f-m62b2\" (UID: \"7718b7b0-f1a7-4b87-8ebe-345e9100463d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-m62b2" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.511104 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7bvj\" (UniqueName: \"kubernetes.io/projected/dab28707-1fc8-45a4-8ad4-0842b7f70239-kube-api-access-t7bvj\") pod \"console-operator-58897d9998-ct5l6\" (UID: \"dab28707-1fc8-45a4-8ad4-0842b7f70239\") " pod="openshift-console-operator/console-operator-58897d9998-ct5l6" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.511124 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7718b7b0-f1a7-4b87-8ebe-345e9100463d-serving-cert\") pod \"controller-manager-879f6c89f-m62b2\" (UID: \"7718b7b0-f1a7-4b87-8ebe-345e9100463d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-m62b2" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.511146 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a16e3671-5260-4e88-8fcc-8ba2252af6ea-config\") pod \"authentication-operator-69f744f599-mt8x9\" (UID: \"a16e3671-5260-4e88-8fcc-8ba2252af6ea\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mt8x9" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.511182 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a16e3671-5260-4e88-8fcc-8ba2252af6ea-serving-cert\") pod \"authentication-operator-69f744f599-mt8x9\" (UID: \"a16e3671-5260-4e88-8fcc-8ba2252af6ea\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mt8x9" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.511204 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-4cq4j\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " pod="openshift-authentication/oauth-openshift-558db77b4-4cq4j" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.511215 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-f8l78" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.511234 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-4cq4j\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " pod="openshift-authentication/oauth-openshift-558db77b4-4cq4j" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.511254 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a16e3671-5260-4e88-8fcc-8ba2252af6ea-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-mt8x9\" (UID: \"a16e3671-5260-4e88-8fcc-8ba2252af6ea\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mt8x9" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.511310 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-4cq4j\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " pod="openshift-authentication/oauth-openshift-558db77b4-4cq4j" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.511331 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8fddc32-a0a6-4149-aa39-b7f51f78a772-config\") pod \"machine-approver-56656f9798-79hfl\" (UID: \"d8fddc32-a0a6-4149-aa39-b7f51f78a772\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-79hfl" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.511352 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7718b7b0-f1a7-4b87-8ebe-345e9100463d-client-ca\") pod \"controller-manager-879f6c89f-m62b2\" (UID: \"7718b7b0-f1a7-4b87-8ebe-345e9100463d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-m62b2" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.511397 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d667dd66-2215-4d9c-9cf5-64ea2f9a2dc8-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-r8cnf\" (UID: \"d667dd66-2215-4d9c-9cf5-64ea2f9a2dc8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r8cnf" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.511417 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/b9dce599-89b8-4c6c-a12c-65f9ef64de9b-audit\") pod \"apiserver-76f77b778f-wd9qj\" (UID: \"b9dce599-89b8-4c6c-a12c-65f9ef64de9b\") " pod="openshift-apiserver/apiserver-76f77b778f-wd9qj" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.511437 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dab28707-1fc8-45a4-8ad4-0842b7f70239-serving-cert\") pod \"console-operator-58897d9998-ct5l6\" (UID: \"dab28707-1fc8-45a4-8ad4-0842b7f70239\") " pod="openshift-console-operator/console-operator-58897d9998-ct5l6" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.511457 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/b9dce599-89b8-4c6c-a12c-65f9ef64de9b-etcd-serving-ca\") pod \"apiserver-76f77b778f-wd9qj\" (UID: \"b9dce599-89b8-4c6c-a12c-65f9ef64de9b\") " pod="openshift-apiserver/apiserver-76f77b778f-wd9qj" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.511476 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-4cq4j\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " pod="openshift-authentication/oauth-openshift-558db77b4-4cq4j" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.511495 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b9dce599-89b8-4c6c-a12c-65f9ef64de9b-etcd-client\") pod \"apiserver-76f77b778f-wd9qj\" (UID: \"b9dce599-89b8-4c6c-a12c-65f9ef64de9b\") " pod="openshift-apiserver/apiserver-76f77b778f-wd9qj" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.511518 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-4cq4j\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " pod="openshift-authentication/oauth-openshift-558db77b4-4cq4j" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.511542 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b9dce599-89b8-4c6c-a12c-65f9ef64de9b-trusted-ca-bundle\") pod \"apiserver-76f77b778f-wd9qj\" (UID: \"b9dce599-89b8-4c6c-a12c-65f9ef64de9b\") " pod="openshift-apiserver/apiserver-76f77b778f-wd9qj" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.511561 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfvl2\" (UniqueName: \"kubernetes.io/projected/7718b7b0-f1a7-4b87-8ebe-345e9100463d-kube-api-access-vfvl2\") pod \"controller-manager-879f6c89f-m62b2\" (UID: \"7718b7b0-f1a7-4b87-8ebe-345e9100463d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-m62b2" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.511583 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-4cq4j\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " pod="openshift-authentication/oauth-openshift-558db77b4-4cq4j" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.511605 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7718b7b0-f1a7-4b87-8ebe-345e9100463d-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-m62b2\" (UID: \"7718b7b0-f1a7-4b87-8ebe-345e9100463d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-m62b2" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.511622 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d8fddc32-a0a6-4149-aa39-b7f51f78a772-auth-proxy-config\") pod \"machine-approver-56656f9798-79hfl\" (UID: \"d8fddc32-a0a6-4149-aa39-b7f51f78a772\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-79hfl" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.511641 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbxp6\" (UniqueName: \"kubernetes.io/projected/d8fddc32-a0a6-4149-aa39-b7f51f78a772-kube-api-access-rbxp6\") pod \"machine-approver-56656f9798-79hfl\" (UID: \"d8fddc32-a0a6-4149-aa39-b7f51f78a772\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-79hfl" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.511659 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/b9dce599-89b8-4c6c-a12c-65f9ef64de9b-node-pullsecrets\") pod \"apiserver-76f77b778f-wd9qj\" (UID: \"b9dce599-89b8-4c6c-a12c-65f9ef64de9b\") " pod="openshift-apiserver/apiserver-76f77b778f-wd9qj" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.511680 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pppg7\" (UniqueName: \"kubernetes.io/projected/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-kube-api-access-pppg7\") pod \"oauth-openshift-558db77b4-4cq4j\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " pod="openshift-authentication/oauth-openshift-558db77b4-4cq4j" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.511698 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d667dd66-2215-4d9c-9cf5-64ea2f9a2dc8-config\") pod \"openshift-apiserver-operator-796bbdcf4f-r8cnf\" (UID: \"d667dd66-2215-4d9c-9cf5-64ea2f9a2dc8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r8cnf" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.511718 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ztcgq\" (UniqueName: \"kubernetes.io/projected/d667dd66-2215-4d9c-9cf5-64ea2f9a2dc8-kube-api-access-ztcgq\") pod \"openshift-apiserver-operator-796bbdcf4f-r8cnf\" (UID: \"d667dd66-2215-4d9c-9cf5-64ea2f9a2dc8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r8cnf" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.511738 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a16e3671-5260-4e88-8fcc-8ba2252af6ea-service-ca-bundle\") pod \"authentication-operator-69f744f599-mt8x9\" (UID: \"a16e3671-5260-4e88-8fcc-8ba2252af6ea\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mt8x9" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.511759 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-4cq4j\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " pod="openshift-authentication/oauth-openshift-558db77b4-4cq4j" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.511784 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9dce599-89b8-4c6c-a12c-65f9ef64de9b-config\") pod \"apiserver-76f77b778f-wd9qj\" (UID: \"b9dce599-89b8-4c6c-a12c-65f9ef64de9b\") " pod="openshift-apiserver/apiserver-76f77b778f-wd9qj" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.511819 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9w6sz\" (UniqueName: \"kubernetes.io/projected/a3e78085-1f5d-4f0c-84bb-62b787b1fbdd-kube-api-access-9w6sz\") pod \"machine-api-operator-5694c8668f-74k4s\" (UID: \"a3e78085-1f5d-4f0c-84bb-62b787b1fbdd\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-74k4s" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.511838 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/b9dce599-89b8-4c6c-a12c-65f9ef64de9b-image-import-ca\") pod \"apiserver-76f77b778f-wd9qj\" (UID: \"b9dce599-89b8-4c6c-a12c-65f9ef64de9b\") " pod="openshift-apiserver/apiserver-76f77b778f-wd9qj" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.511856 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b9dce599-89b8-4c6c-a12c-65f9ef64de9b-audit-dir\") pod \"apiserver-76f77b778f-wd9qj\" (UID: \"b9dce599-89b8-4c6c-a12c-65f9ef64de9b\") " pod="openshift-apiserver/apiserver-76f77b778f-wd9qj" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.511876 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3e78085-1f5d-4f0c-84bb-62b787b1fbdd-config\") pod \"machine-api-operator-5694c8668f-74k4s\" (UID: \"a3e78085-1f5d-4f0c-84bb-62b787b1fbdd\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-74k4s" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.511895 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-4cq4j\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " pod="openshift-authentication/oauth-openshift-558db77b4-4cq4j" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.514808 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.515119 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-l5vld"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.515940 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hkkwq"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.516501 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hkkwq" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.518715 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.519310 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l5vld" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.529901 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.534275 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.534817 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.535049 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.536978 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.537154 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.537460 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.537614 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.537737 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.537883 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.538185 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.538447 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.552457 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tdwfd"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.553264 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tdwfd" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.571050 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.571355 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.571641 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.571797 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.572076 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.573450 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rdfcg"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.574056 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-gm548"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.574584 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gm548" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.575000 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rdfcg" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.575845 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-r76vv"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.576683 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.576805 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-r76vv" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.576885 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.577033 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.577552 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.577746 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.578427 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.578759 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.578949 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.579050 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.579624 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.579731 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.583241 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.590829 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pj8qs"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.591455 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pj8qs" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.591825 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.594238 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.594294 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.594558 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.597714 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dkktg"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.598147 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dkktg" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.610845 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.611077 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.611263 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.611446 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.611518 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.612522 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.612615 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.612744 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.612818 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.612883 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.612954 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.613020 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.613145 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.613196 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.613275 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.613332 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.613353 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.613433 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.613458 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.613148 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.613519 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.613582 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.613654 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.613685 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.613722 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.613762 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.613839 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.613852 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.613906 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.613916 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.613980 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.614044 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.614058 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.614125 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.614187 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.614541 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a16e3671-5260-4e88-8fcc-8ba2252af6ea-config\") pod \"authentication-operator-69f744f599-mt8x9\" (UID: \"a16e3671-5260-4e88-8fcc-8ba2252af6ea\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mt8x9" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.614568 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7bvj\" (UniqueName: \"kubernetes.io/projected/dab28707-1fc8-45a4-8ad4-0842b7f70239-kube-api-access-t7bvj\") pod \"console-operator-58897d9998-ct5l6\" (UID: \"dab28707-1fc8-45a4-8ad4-0842b7f70239\") " pod="openshift-console-operator/console-operator-58897d9998-ct5l6" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.614585 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7718b7b0-f1a7-4b87-8ebe-345e9100463d-serving-cert\") pod \"controller-manager-879f6c89f-m62b2\" (UID: \"7718b7b0-f1a7-4b87-8ebe-345e9100463d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-m62b2" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.614605 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/745ab1fe-8d37-420f-9c12-62cd7a33aeaa-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-hkkwq\" (UID: \"745ab1fe-8d37-420f-9c12-62cd7a33aeaa\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hkkwq" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.614622 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2b42593e-4396-4b1a-9272-ca2a2fb5ad6e-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-l5vld\" (UID: \"2b42593e-4396-4b1a-9272-ca2a2fb5ad6e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l5vld" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.614637 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqq2z\" (UniqueName: \"kubernetes.io/projected/bd697754-4524-44ed-be19-18747c0626f5-kube-api-access-qqq2z\") pod \"control-plane-machine-set-operator-78cbb6b69f-dkktg\" (UID: \"bd697754-4524-44ed-be19-18747c0626f5\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dkktg" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.614655 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a16e3671-5260-4e88-8fcc-8ba2252af6ea-serving-cert\") pod \"authentication-operator-69f744f599-mt8x9\" (UID: \"a16e3671-5260-4e88-8fcc-8ba2252af6ea\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mt8x9" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.614670 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c2b2606-ba50-42a8-ae0b-2176b853cb94-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-tdwfd\" (UID: \"3c2b2606-ba50-42a8-ae0b-2176b853cb94\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tdwfd" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.614684 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggbks\" (UniqueName: \"kubernetes.io/projected/3c2b2606-ba50-42a8-ae0b-2176b853cb94-kube-api-access-ggbks\") pod \"openshift-controller-manager-operator-756b6f6bc6-tdwfd\" (UID: \"3c2b2606-ba50-42a8-ae0b-2176b853cb94\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tdwfd" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.614699 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/74e9c10d-c66e-48ea-9813-cb221012dda4-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-pj8qs\" (UID: \"74e9c10d-c66e-48ea-9813-cb221012dda4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pj8qs" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.614713 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4e2257ac-9d7a-4e81-b560-32f28a73830e-metrics-tls\") pod \"dns-operator-744455d44c-b68qv\" (UID: \"4e2257ac-9d7a-4e81-b560-32f28a73830e\") " pod="openshift-dns-operator/dns-operator-744455d44c-b68qv" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.614727 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2b42593e-4396-4b1a-9272-ca2a2fb5ad6e-audit-dir\") pod \"apiserver-7bbb656c7d-l5vld\" (UID: \"2b42593e-4396-4b1a-9272-ca2a2fb5ad6e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l5vld" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.614744 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-4cq4j\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " pod="openshift-authentication/oauth-openshift-558db77b4-4cq4j" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.614759 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7829943-cabc-49c5-a547-62124e38d49e-config\") pod \"route-controller-manager-6576b87f9c-98h6l\" (UID: \"f7829943-cabc-49c5-a547-62124e38d49e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-98h6l" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.614783 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-4cq4j\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " pod="openshift-authentication/oauth-openshift-558db77b4-4cq4j" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.614798 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74e9c10d-c66e-48ea-9813-cb221012dda4-config\") pod \"kube-apiserver-operator-766d6c64bb-pj8qs\" (UID: \"74e9c10d-c66e-48ea-9813-cb221012dda4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pj8qs" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.614813 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a00e530f-3c79-40da-a3f3-faf504af3fea-serving-cert\") pod \"openshift-config-operator-7777fb866f-gm548\" (UID: \"a00e530f-3c79-40da-a3f3-faf504af3fea\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gm548" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.614828 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a16e3671-5260-4e88-8fcc-8ba2252af6ea-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-mt8x9\" (UID: \"a16e3671-5260-4e88-8fcc-8ba2252af6ea\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mt8x9" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.614844 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-4cq4j\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " pod="openshift-authentication/oauth-openshift-558db77b4-4cq4j" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.614858 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3c2b2606-ba50-42a8-ae0b-2176b853cb94-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-tdwfd\" (UID: \"3c2b2606-ba50-42a8-ae0b-2176b853cb94\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tdwfd" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.614873 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7829943-cabc-49c5-a547-62124e38d49e-serving-cert\") pod \"route-controller-manager-6576b87f9c-98h6l\" (UID: \"f7829943-cabc-49c5-a547-62124e38d49e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-98h6l" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.614889 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8fddc32-a0a6-4149-aa39-b7f51f78a772-config\") pod \"machine-approver-56656f9798-79hfl\" (UID: \"d8fddc32-a0a6-4149-aa39-b7f51f78a772\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-79hfl" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.614903 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7718b7b0-f1a7-4b87-8ebe-345e9100463d-client-ca\") pod \"controller-manager-879f6c89f-m62b2\" (UID: \"7718b7b0-f1a7-4b87-8ebe-345e9100463d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-m62b2" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.614919 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/a00e530f-3c79-40da-a3f3-faf504af3fea-available-featuregates\") pod \"openshift-config-operator-7777fb866f-gm548\" (UID: \"a00e530f-3c79-40da-a3f3-faf504af3fea\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gm548" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.614941 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d667dd66-2215-4d9c-9cf5-64ea2f9a2dc8-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-r8cnf\" (UID: \"d667dd66-2215-4d9c-9cf5-64ea2f9a2dc8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r8cnf" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.614957 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/b9dce599-89b8-4c6c-a12c-65f9ef64de9b-audit\") pod \"apiserver-76f77b778f-wd9qj\" (UID: \"b9dce599-89b8-4c6c-a12c-65f9ef64de9b\") " pod="openshift-apiserver/apiserver-76f77b778f-wd9qj" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.614972 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dab28707-1fc8-45a4-8ad4-0842b7f70239-serving-cert\") pod \"console-operator-58897d9998-ct5l6\" (UID: \"dab28707-1fc8-45a4-8ad4-0842b7f70239\") " pod="openshift-console-operator/console-operator-58897d9998-ct5l6" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.614988 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dw6wr\" (UniqueName: \"kubernetes.io/projected/317b7923-187f-47a4-9b82-53b20c86d9a4-kube-api-access-dw6wr\") pod \"console-f9d7485db-f8l78\" (UID: \"317b7923-187f-47a4-9b82-53b20c86d9a4\") " pod="openshift-console/console-f9d7485db-f8l78" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615002 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xclz\" (UniqueName: \"kubernetes.io/projected/4e2257ac-9d7a-4e81-b560-32f28a73830e-kube-api-access-4xclz\") pod \"dns-operator-744455d44c-b68qv\" (UID: \"4e2257ac-9d7a-4e81-b560-32f28a73830e\") " pod="openshift-dns-operator/dns-operator-744455d44c-b68qv" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615016 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/b9dce599-89b8-4c6c-a12c-65f9ef64de9b-etcd-serving-ca\") pod \"apiserver-76f77b778f-wd9qj\" (UID: \"b9dce599-89b8-4c6c-a12c-65f9ef64de9b\") " pod="openshift-apiserver/apiserver-76f77b778f-wd9qj" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615032 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-4cq4j\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " pod="openshift-authentication/oauth-openshift-558db77b4-4cq4j" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615047 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b9dce599-89b8-4c6c-a12c-65f9ef64de9b-etcd-client\") pod \"apiserver-76f77b778f-wd9qj\" (UID: \"b9dce599-89b8-4c6c-a12c-65f9ef64de9b\") " pod="openshift-apiserver/apiserver-76f77b778f-wd9qj" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615063 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-4cq4j\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " pod="openshift-authentication/oauth-openshift-558db77b4-4cq4j" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615078 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b9dce599-89b8-4c6c-a12c-65f9ef64de9b-trusted-ca-bundle\") pod \"apiserver-76f77b778f-wd9qj\" (UID: \"b9dce599-89b8-4c6c-a12c-65f9ef64de9b\") " pod="openshift-apiserver/apiserver-76f77b778f-wd9qj" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615094 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2b42593e-4396-4b1a-9272-ca2a2fb5ad6e-audit-policies\") pod \"apiserver-7bbb656c7d-l5vld\" (UID: \"2b42593e-4396-4b1a-9272-ca2a2fb5ad6e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l5vld" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615108 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2b42593e-4396-4b1a-9272-ca2a2fb5ad6e-etcd-client\") pod \"apiserver-7bbb656c7d-l5vld\" (UID: \"2b42593e-4396-4b1a-9272-ca2a2fb5ad6e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l5vld" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615125 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfvl2\" (UniqueName: \"kubernetes.io/projected/7718b7b0-f1a7-4b87-8ebe-345e9100463d-kube-api-access-vfvl2\") pod \"controller-manager-879f6c89f-m62b2\" (UID: \"7718b7b0-f1a7-4b87-8ebe-345e9100463d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-m62b2" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615141 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-4cq4j\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " pod="openshift-authentication/oauth-openshift-558db77b4-4cq4j" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615173 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5f8gt\" (UniqueName: \"kubernetes.io/projected/f7829943-cabc-49c5-a547-62124e38d49e-kube-api-access-5f8gt\") pod \"route-controller-manager-6576b87f9c-98h6l\" (UID: \"f7829943-cabc-49c5-a547-62124e38d49e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-98h6l" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615189 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7718b7b0-f1a7-4b87-8ebe-345e9100463d-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-m62b2\" (UID: \"7718b7b0-f1a7-4b87-8ebe-345e9100463d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-m62b2" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615206 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdbpq\" (UniqueName: \"kubernetes.io/projected/745ab1fe-8d37-420f-9c12-62cd7a33aeaa-kube-api-access-mdbpq\") pod \"cluster-image-registry-operator-dc59b4c8b-hkkwq\" (UID: \"745ab1fe-8d37-420f-9c12-62cd7a33aeaa\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hkkwq" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615222 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbxp6\" (UniqueName: \"kubernetes.io/projected/d8fddc32-a0a6-4149-aa39-b7f51f78a772-kube-api-access-rbxp6\") pod \"machine-approver-56656f9798-79hfl\" (UID: \"d8fddc32-a0a6-4149-aa39-b7f51f78a772\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-79hfl" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615238 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/b9dce599-89b8-4c6c-a12c-65f9ef64de9b-node-pullsecrets\") pod \"apiserver-76f77b778f-wd9qj\" (UID: \"b9dce599-89b8-4c6c-a12c-65f9ef64de9b\") " pod="openshift-apiserver/apiserver-76f77b778f-wd9qj" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615254 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6rpm\" (UniqueName: \"kubernetes.io/projected/2b42593e-4396-4b1a-9272-ca2a2fb5ad6e-kube-api-access-k6rpm\") pod \"apiserver-7bbb656c7d-l5vld\" (UID: \"2b42593e-4396-4b1a-9272-ca2a2fb5ad6e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l5vld" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615268 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/bd697754-4524-44ed-be19-18747c0626f5-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-dkktg\" (UID: \"bd697754-4524-44ed-be19-18747c0626f5\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dkktg" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615284 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tw8r\" (UniqueName: \"kubernetes.io/projected/56fd5dd1-12e2-4b22-a4ab-291d6bb3ce92-kube-api-access-2tw8r\") pod \"ingress-operator-5b745b69d9-r76vv\" (UID: \"56fd5dd1-12e2-4b22-a4ab-291d6bb3ce92\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-r76vv" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615299 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d8fddc32-a0a6-4149-aa39-b7f51f78a772-auth-proxy-config\") pod \"machine-approver-56656f9798-79hfl\" (UID: \"d8fddc32-a0a6-4149-aa39-b7f51f78a772\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-79hfl" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615314 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a16e3671-5260-4e88-8fcc-8ba2252af6ea-service-ca-bundle\") pod \"authentication-operator-69f744f599-mt8x9\" (UID: \"a16e3671-5260-4e88-8fcc-8ba2252af6ea\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mt8x9" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615352 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pppg7\" (UniqueName: \"kubernetes.io/projected/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-kube-api-access-pppg7\") pod \"oauth-openshift-558db77b4-4cq4j\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " pod="openshift-authentication/oauth-openshift-558db77b4-4cq4j" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615369 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d667dd66-2215-4d9c-9cf5-64ea2f9a2dc8-config\") pod \"openshift-apiserver-operator-796bbdcf4f-r8cnf\" (UID: \"d667dd66-2215-4d9c-9cf5-64ea2f9a2dc8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r8cnf" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615383 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ztcgq\" (UniqueName: \"kubernetes.io/projected/d667dd66-2215-4d9c-9cf5-64ea2f9a2dc8-kube-api-access-ztcgq\") pod \"openshift-apiserver-operator-796bbdcf4f-r8cnf\" (UID: \"d667dd66-2215-4d9c-9cf5-64ea2f9a2dc8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r8cnf" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615398 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/56fd5dd1-12e2-4b22-a4ab-291d6bb3ce92-trusted-ca\") pod \"ingress-operator-5b745b69d9-r76vv\" (UID: \"56fd5dd1-12e2-4b22-a4ab-291d6bb3ce92\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-r76vv" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615415 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-4cq4j\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " pod="openshift-authentication/oauth-openshift-558db77b4-4cq4j" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615429 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9dce599-89b8-4c6c-a12c-65f9ef64de9b-config\") pod \"apiserver-76f77b778f-wd9qj\" (UID: \"b9dce599-89b8-4c6c-a12c-65f9ef64de9b\") " pod="openshift-apiserver/apiserver-76f77b778f-wd9qj" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615445 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/745ab1fe-8d37-420f-9c12-62cd7a33aeaa-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-hkkwq\" (UID: \"745ab1fe-8d37-420f-9c12-62cd7a33aeaa\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hkkwq" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615459 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/317b7923-187f-47a4-9b82-53b20c86d9a4-service-ca\") pod \"console-f9d7485db-f8l78\" (UID: \"317b7923-187f-47a4-9b82-53b20c86d9a4\") " pod="openshift-console/console-f9d7485db-f8l78" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615482 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/317b7923-187f-47a4-9b82-53b20c86d9a4-console-oauth-config\") pod \"console-f9d7485db-f8l78\" (UID: \"317b7923-187f-47a4-9b82-53b20c86d9a4\") " pod="openshift-console/console-f9d7485db-f8l78" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615497 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/317b7923-187f-47a4-9b82-53b20c86d9a4-trusted-ca-bundle\") pod \"console-f9d7485db-f8l78\" (UID: \"317b7923-187f-47a4-9b82-53b20c86d9a4\") " pod="openshift-console/console-f9d7485db-f8l78" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615516 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9w6sz\" (UniqueName: \"kubernetes.io/projected/a3e78085-1f5d-4f0c-84bb-62b787b1fbdd-kube-api-access-9w6sz\") pod \"machine-api-operator-5694c8668f-74k4s\" (UID: \"a3e78085-1f5d-4f0c-84bb-62b787b1fbdd\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-74k4s" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615532 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/b9dce599-89b8-4c6c-a12c-65f9ef64de9b-image-import-ca\") pod \"apiserver-76f77b778f-wd9qj\" (UID: \"b9dce599-89b8-4c6c-a12c-65f9ef64de9b\") " pod="openshift-apiserver/apiserver-76f77b778f-wd9qj" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615547 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b9dce599-89b8-4c6c-a12c-65f9ef64de9b-audit-dir\") pod \"apiserver-76f77b778f-wd9qj\" (UID: \"b9dce599-89b8-4c6c-a12c-65f9ef64de9b\") " pod="openshift-apiserver/apiserver-76f77b778f-wd9qj" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615563 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/74e9c10d-c66e-48ea-9813-cb221012dda4-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-pj8qs\" (UID: \"74e9c10d-c66e-48ea-9813-cb221012dda4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pj8qs" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615579 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3e78085-1f5d-4f0c-84bb-62b787b1fbdd-config\") pod \"machine-api-operator-5694c8668f-74k4s\" (UID: \"a3e78085-1f5d-4f0c-84bb-62b787b1fbdd\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-74k4s" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615595 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-4cq4j\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " pod="openshift-authentication/oauth-openshift-558db77b4-4cq4j" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615609 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/4849186f-21ae-4654-a813-4b176093abbd-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-rdfcg\" (UID: \"4849186f-21ae-4654-a813-4b176093abbd\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rdfcg" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615624 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f7829943-cabc-49c5-a547-62124e38d49e-client-ca\") pod \"route-controller-manager-6576b87f9c-98h6l\" (UID: \"f7829943-cabc-49c5-a547-62124e38d49e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-98h6l" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615641 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/a3e78085-1f5d-4f0c-84bb-62b787b1fbdd-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-74k4s\" (UID: \"a3e78085-1f5d-4f0c-84bb-62b787b1fbdd\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-74k4s" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615656 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dab28707-1fc8-45a4-8ad4-0842b7f70239-trusted-ca\") pod \"console-operator-58897d9998-ct5l6\" (UID: \"dab28707-1fc8-45a4-8ad4-0842b7f70239\") " pod="openshift-console-operator/console-operator-58897d9998-ct5l6" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615672 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/d8fddc32-a0a6-4149-aa39-b7f51f78a772-machine-approver-tls\") pod \"machine-approver-56656f9798-79hfl\" (UID: \"d8fddc32-a0a6-4149-aa39-b7f51f78a772\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-79hfl" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615687 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-4cq4j\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " pod="openshift-authentication/oauth-openshift-558db77b4-4cq4j" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615702 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2b42593e-4396-4b1a-9272-ca2a2fb5ad6e-encryption-config\") pod \"apiserver-7bbb656c7d-l5vld\" (UID: \"2b42593e-4396-4b1a-9272-ca2a2fb5ad6e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l5vld" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615725 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-4cq4j\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " pod="openshift-authentication/oauth-openshift-558db77b4-4cq4j" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615741 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/b9dce599-89b8-4c6c-a12c-65f9ef64de9b-encryption-config\") pod \"apiserver-76f77b778f-wd9qj\" (UID: \"b9dce599-89b8-4c6c-a12c-65f9ef64de9b\") " pod="openshift-apiserver/apiserver-76f77b778f-wd9qj" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615757 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzwg2\" (UniqueName: \"kubernetes.io/projected/b9dce599-89b8-4c6c-a12c-65f9ef64de9b-kube-api-access-lzwg2\") pod \"apiserver-76f77b778f-wd9qj\" (UID: \"b9dce599-89b8-4c6c-a12c-65f9ef64de9b\") " pod="openshift-apiserver/apiserver-76f77b778f-wd9qj" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615773 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dab28707-1fc8-45a4-8ad4-0842b7f70239-config\") pod \"console-operator-58897d9998-ct5l6\" (UID: \"dab28707-1fc8-45a4-8ad4-0842b7f70239\") " pod="openshift-console-operator/console-operator-58897d9998-ct5l6" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615789 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hphpt\" (UniqueName: \"kubernetes.io/projected/a00e530f-3c79-40da-a3f3-faf504af3fea-kube-api-access-hphpt\") pod \"openshift-config-operator-7777fb866f-gm548\" (UID: \"a00e530f-3c79-40da-a3f3-faf504af3fea\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gm548" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615806 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/317b7923-187f-47a4-9b82-53b20c86d9a4-oauth-serving-cert\") pod \"console-f9d7485db-f8l78\" (UID: \"317b7923-187f-47a4-9b82-53b20c86d9a4\") " pod="openshift-console/console-f9d7485db-f8l78" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615820 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/317b7923-187f-47a4-9b82-53b20c86d9a4-console-serving-cert\") pod \"console-f9d7485db-f8l78\" (UID: \"317b7923-187f-47a4-9b82-53b20c86d9a4\") " pod="openshift-console/console-f9d7485db-f8l78" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615836 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-audit-policies\") pod \"oauth-openshift-558db77b4-4cq4j\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " pod="openshift-authentication/oauth-openshift-558db77b4-4cq4j" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615851 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/56fd5dd1-12e2-4b22-a4ab-291d6bb3ce92-bound-sa-token\") pod \"ingress-operator-5b745b69d9-r76vv\" (UID: \"56fd5dd1-12e2-4b22-a4ab-291d6bb3ce92\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-r76vv" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615867 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b9dce599-89b8-4c6c-a12c-65f9ef64de9b-serving-cert\") pod \"apiserver-76f77b778f-wd9qj\" (UID: \"b9dce599-89b8-4c6c-a12c-65f9ef64de9b\") " pod="openshift-apiserver/apiserver-76f77b778f-wd9qj" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615881 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/317b7923-187f-47a4-9b82-53b20c86d9a4-console-config\") pod \"console-f9d7485db-f8l78\" (UID: \"317b7923-187f-47a4-9b82-53b20c86d9a4\") " pod="openshift-console/console-f9d7485db-f8l78" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615907 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/a3e78085-1f5d-4f0c-84bb-62b787b1fbdd-images\") pod \"machine-api-operator-5694c8668f-74k4s\" (UID: \"a3e78085-1f5d-4f0c-84bb-62b787b1fbdd\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-74k4s" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615922 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-audit-dir\") pod \"oauth-openshift-558db77b4-4cq4j\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " pod="openshift-authentication/oauth-openshift-558db77b4-4cq4j" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615938 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/745ab1fe-8d37-420f-9c12-62cd7a33aeaa-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-hkkwq\" (UID: \"745ab1fe-8d37-420f-9c12-62cd7a33aeaa\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hkkwq" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615953 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgzxc\" (UniqueName: \"kubernetes.io/projected/4849186f-21ae-4654-a813-4b176093abbd-kube-api-access-pgzxc\") pod \"cluster-samples-operator-665b6dd947-rdfcg\" (UID: \"4849186f-21ae-4654-a813-4b176093abbd\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rdfcg" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615972 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8c5w\" (UniqueName: \"kubernetes.io/projected/a16e3671-5260-4e88-8fcc-8ba2252af6ea-kube-api-access-g8c5w\") pod \"authentication-operator-69f744f599-mt8x9\" (UID: \"a16e3671-5260-4e88-8fcc-8ba2252af6ea\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mt8x9" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.615986 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-4cq4j\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " pod="openshift-authentication/oauth-openshift-558db77b4-4cq4j" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.616001 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7718b7b0-f1a7-4b87-8ebe-345e9100463d-config\") pod \"controller-manager-879f6c89f-m62b2\" (UID: \"7718b7b0-f1a7-4b87-8ebe-345e9100463d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-m62b2" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.616023 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b42593e-4396-4b1a-9272-ca2a2fb5ad6e-serving-cert\") pod \"apiserver-7bbb656c7d-l5vld\" (UID: \"2b42593e-4396-4b1a-9272-ca2a2fb5ad6e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l5vld" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.616040 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/56fd5dd1-12e2-4b22-a4ab-291d6bb3ce92-metrics-tls\") pod \"ingress-operator-5b745b69d9-r76vv\" (UID: \"56fd5dd1-12e2-4b22-a4ab-291d6bb3ce92\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-r76vv" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.616054 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2b42593e-4396-4b1a-9272-ca2a2fb5ad6e-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-l5vld\" (UID: \"2b42593e-4396-4b1a-9272-ca2a2fb5ad6e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l5vld" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.625178 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d8fddc32-a0a6-4149-aa39-b7f51f78a772-auth-proxy-config\") pod \"machine-approver-56656f9798-79hfl\" (UID: \"d8fddc32-a0a6-4149-aa39-b7f51f78a772\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-79hfl" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.625361 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a16e3671-5260-4e88-8fcc-8ba2252af6ea-config\") pod \"authentication-operator-69f744f599-mt8x9\" (UID: \"a16e3671-5260-4e88-8fcc-8ba2252af6ea\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mt8x9" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.627111 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/b9dce599-89b8-4c6c-a12c-65f9ef64de9b-etcd-serving-ca\") pod \"apiserver-76f77b778f-wd9qj\" (UID: \"b9dce599-89b8-4c6c-a12c-65f9ef64de9b\") " pod="openshift-apiserver/apiserver-76f77b778f-wd9qj" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.628060 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/b9dce599-89b8-4c6c-a12c-65f9ef64de9b-audit\") pod \"apiserver-76f77b778f-wd9qj\" (UID: \"b9dce599-89b8-4c6c-a12c-65f9ef64de9b\") " pod="openshift-apiserver/apiserver-76f77b778f-wd9qj" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.629205 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.685083 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-4cq4j\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " pod="openshift-authentication/oauth-openshift-558db77b4-4cq4j" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.631182 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b9dce599-89b8-4c6c-a12c-65f9ef64de9b-audit-dir\") pod \"apiserver-76f77b778f-wd9qj\" (UID: \"b9dce599-89b8-4c6c-a12c-65f9ef64de9b\") " pod="openshift-apiserver/apiserver-76f77b778f-wd9qj" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.631808 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-4cq4j\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " pod="openshift-authentication/oauth-openshift-558db77b4-4cq4j" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.632293 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3e78085-1f5d-4f0c-84bb-62b787b1fbdd-config\") pod \"machine-api-operator-5694c8668f-74k4s\" (UID: \"a3e78085-1f5d-4f0c-84bb-62b787b1fbdd\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-74k4s" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.631136 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/b9dce599-89b8-4c6c-a12c-65f9ef64de9b-image-import-ca\") pod \"apiserver-76f77b778f-wd9qj\" (UID: \"b9dce599-89b8-4c6c-a12c-65f9ef64de9b\") " pod="openshift-apiserver/apiserver-76f77b778f-wd9qj" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.637860 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-4cq4j\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " pod="openshift-authentication/oauth-openshift-558db77b4-4cq4j" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.638191 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-4cq4j\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " pod="openshift-authentication/oauth-openshift-558db77b4-4cq4j" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.636350 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dab28707-1fc8-45a4-8ad4-0842b7f70239-serving-cert\") pod \"console-operator-58897d9998-ct5l6\" (UID: \"dab28707-1fc8-45a4-8ad4-0842b7f70239\") " pod="openshift-console-operator/console-operator-58897d9998-ct5l6" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.638371 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8fddc32-a0a6-4149-aa39-b7f51f78a772-config\") pod \"machine-approver-56656f9798-79hfl\" (UID: \"d8fddc32-a0a6-4149-aa39-b7f51f78a772\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-79hfl" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.638484 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/d8fddc32-a0a6-4149-aa39-b7f51f78a772-machine-approver-tls\") pod \"machine-approver-56656f9798-79hfl\" (UID: \"d8fddc32-a0a6-4149-aa39-b7f51f78a772\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-79hfl" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.639114 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9dce599-89b8-4c6c-a12c-65f9ef64de9b-config\") pod \"apiserver-76f77b778f-wd9qj\" (UID: \"b9dce599-89b8-4c6c-a12c-65f9ef64de9b\") " pod="openshift-apiserver/apiserver-76f77b778f-wd9qj" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.639483 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a16e3671-5260-4e88-8fcc-8ba2252af6ea-serving-cert\") pod \"authentication-operator-69f744f599-mt8x9\" (UID: \"a16e3671-5260-4e88-8fcc-8ba2252af6ea\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mt8x9" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.639524 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7718b7b0-f1a7-4b87-8ebe-345e9100463d-client-ca\") pod \"controller-manager-879f6c89f-m62b2\" (UID: \"7718b7b0-f1a7-4b87-8ebe-345e9100463d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-m62b2" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.639620 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/b9dce599-89b8-4c6c-a12c-65f9ef64de9b-node-pullsecrets\") pod \"apiserver-76f77b778f-wd9qj\" (UID: \"b9dce599-89b8-4c6c-a12c-65f9ef64de9b\") " pod="openshift-apiserver/apiserver-76f77b778f-wd9qj" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.643072 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/a3e78085-1f5d-4f0c-84bb-62b787b1fbdd-images\") pod \"machine-api-operator-5694c8668f-74k4s\" (UID: \"a3e78085-1f5d-4f0c-84bb-62b787b1fbdd\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-74k4s" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.643102 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-audit-dir\") pod \"oauth-openshift-558db77b4-4cq4j\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " pod="openshift-authentication/oauth-openshift-558db77b4-4cq4j" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.644646 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b9dce599-89b8-4c6c-a12c-65f9ef64de9b-etcd-client\") pod \"apiserver-76f77b778f-wd9qj\" (UID: \"b9dce599-89b8-4c6c-a12c-65f9ef64de9b\") " pod="openshift-apiserver/apiserver-76f77b778f-wd9qj" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.645808 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dab28707-1fc8-45a4-8ad4-0842b7f70239-config\") pod \"console-operator-58897d9998-ct5l6\" (UID: \"dab28707-1fc8-45a4-8ad4-0842b7f70239\") " pod="openshift-console-operator/console-operator-58897d9998-ct5l6" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.643708 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d667dd66-2215-4d9c-9cf5-64ea2f9a2dc8-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-r8cnf\" (UID: \"d667dd66-2215-4d9c-9cf5-64ea2f9a2dc8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r8cnf" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.646291 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-audit-policies\") pod \"oauth-openshift-558db77b4-4cq4j\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " pod="openshift-authentication/oauth-openshift-558db77b4-4cq4j" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.650133 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d667dd66-2215-4d9c-9cf5-64ea2f9a2dc8-config\") pod \"openshift-apiserver-operator-796bbdcf4f-r8cnf\" (UID: \"d667dd66-2215-4d9c-9cf5-64ea2f9a2dc8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r8cnf" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.650460 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b9dce599-89b8-4c6c-a12c-65f9ef64de9b-serving-cert\") pod \"apiserver-76f77b778f-wd9qj\" (UID: \"b9dce599-89b8-4c6c-a12c-65f9ef64de9b\") " pod="openshift-apiserver/apiserver-76f77b778f-wd9qj" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.647812 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-4cq4j\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " pod="openshift-authentication/oauth-openshift-558db77b4-4cq4j" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.651535 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-hpgjp"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.685786 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-4cq4j\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " pod="openshift-authentication/oauth-openshift-558db77b4-4cq4j" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.654543 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/a3e78085-1f5d-4f0c-84bb-62b787b1fbdd-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-74k4s\" (UID: \"a3e78085-1f5d-4f0c-84bb-62b787b1fbdd\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-74k4s" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.654795 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-4cq4j\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " pod="openshift-authentication/oauth-openshift-558db77b4-4cq4j" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.684657 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-4cq4j\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " pod="openshift-authentication/oauth-openshift-558db77b4-4cq4j" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.651516 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7718b7b0-f1a7-4b87-8ebe-345e9100463d-config\") pod \"controller-manager-879f6c89f-m62b2\" (UID: \"7718b7b0-f1a7-4b87-8ebe-345e9100463d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-m62b2" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.686397 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a16e3671-5260-4e88-8fcc-8ba2252af6ea-service-ca-bundle\") pod \"authentication-operator-69f744f599-mt8x9\" (UID: \"a16e3671-5260-4e88-8fcc-8ba2252af6ea\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mt8x9" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.688027 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/b9dce599-89b8-4c6c-a12c-65f9ef64de9b-encryption-config\") pod \"apiserver-76f77b778f-wd9qj\" (UID: \"b9dce599-89b8-4c6c-a12c-65f9ef64de9b\") " pod="openshift-apiserver/apiserver-76f77b778f-wd9qj" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.688761 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.691017 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-4cq4j\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " pod="openshift-authentication/oauth-openshift-558db77b4-4cq4j" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.632811 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.646464 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.646558 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.646635 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.649335 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.650510 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.651677 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.652610 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.652855 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.652992 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.653925 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.654198 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7718b7b0-f1a7-4b87-8ebe-345e9100463d-serving-cert\") pod \"controller-manager-879f6c89f-m62b2\" (UID: \"7718b7b0-f1a7-4b87-8ebe-345e9100463d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-m62b2" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.701477 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.718450 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.718598 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.720449 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dw6wr\" (UniqueName: \"kubernetes.io/projected/317b7923-187f-47a4-9b82-53b20c86d9a4-kube-api-access-dw6wr\") pod \"console-f9d7485db-f8l78\" (UID: \"317b7923-187f-47a4-9b82-53b20c86d9a4\") " pod="openshift-console/console-f9d7485db-f8l78" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.720495 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xclz\" (UniqueName: \"kubernetes.io/projected/4e2257ac-9d7a-4e81-b560-32f28a73830e-kube-api-access-4xclz\") pod \"dns-operator-744455d44c-b68qv\" (UID: \"4e2257ac-9d7a-4e81-b560-32f28a73830e\") " pod="openshift-dns-operator/dns-operator-744455d44c-b68qv" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.720524 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2b42593e-4396-4b1a-9272-ca2a2fb5ad6e-audit-policies\") pod \"apiserver-7bbb656c7d-l5vld\" (UID: \"2b42593e-4396-4b1a-9272-ca2a2fb5ad6e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l5vld" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.720545 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2b42593e-4396-4b1a-9272-ca2a2fb5ad6e-etcd-client\") pod \"apiserver-7bbb656c7d-l5vld\" (UID: \"2b42593e-4396-4b1a-9272-ca2a2fb5ad6e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l5vld" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.720667 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xhg5n"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.721077 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.721335 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-mq75m"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.721498 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dab28707-1fc8-45a4-8ad4-0842b7f70239-trusted-ca\") pod \"console-operator-58897d9998-ct5l6\" (UID: \"dab28707-1fc8-45a4-8ad4-0842b7f70239\") " pod="openshift-console-operator/console-operator-58897d9998-ct5l6" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.721590 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.721676 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2b42593e-4396-4b1a-9272-ca2a2fb5ad6e-audit-policies\") pod \"apiserver-7bbb656c7d-l5vld\" (UID: \"2b42593e-4396-4b1a-9272-ca2a2fb5ad6e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l5vld" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.721719 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pmbhw"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.721796 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.720676 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5f8gt\" (UniqueName: \"kubernetes.io/projected/f7829943-cabc-49c5-a547-62124e38d49e-kube-api-access-5f8gt\") pod \"route-controller-manager-6576b87f9c-98h6l\" (UID: \"f7829943-cabc-49c5-a547-62124e38d49e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-98h6l" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.722148 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdbpq\" (UniqueName: \"kubernetes.io/projected/745ab1fe-8d37-420f-9c12-62cd7a33aeaa-kube-api-access-mdbpq\") pod \"cluster-image-registry-operator-dc59b4c8b-hkkwq\" (UID: \"745ab1fe-8d37-420f-9c12-62cd7a33aeaa\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hkkwq" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.722201 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6rpm\" (UniqueName: \"kubernetes.io/projected/2b42593e-4396-4b1a-9272-ca2a2fb5ad6e-kube-api-access-k6rpm\") pod \"apiserver-7bbb656c7d-l5vld\" (UID: \"2b42593e-4396-4b1a-9272-ca2a2fb5ad6e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l5vld" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.722223 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.722332 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-t4x4n"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.722225 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/bd697754-4524-44ed-be19-18747c0626f5-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-dkktg\" (UID: \"bd697754-4524-44ed-be19-18747c0626f5\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dkktg" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.722494 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tw8r\" (UniqueName: \"kubernetes.io/projected/56fd5dd1-12e2-4b22-a4ab-291d6bb3ce92-kube-api-access-2tw8r\") pod \"ingress-operator-5b745b69d9-r76vv\" (UID: \"56fd5dd1-12e2-4b22-a4ab-291d6bb3ce92\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-r76vv" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.722513 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pmbhw" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.722572 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.722554 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/56fd5dd1-12e2-4b22-a4ab-291d6bb3ce92-trusted-ca\") pod \"ingress-operator-5b745b69d9-r76vv\" (UID: \"56fd5dd1-12e2-4b22-a4ab-291d6bb3ce92\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-r76vv" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.722612 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/745ab1fe-8d37-420f-9c12-62cd7a33aeaa-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-hkkwq\" (UID: \"745ab1fe-8d37-420f-9c12-62cd7a33aeaa\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hkkwq" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.722631 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/317b7923-187f-47a4-9b82-53b20c86d9a4-console-oauth-config\") pod \"console-f9d7485db-f8l78\" (UID: \"317b7923-187f-47a4-9b82-53b20c86d9a4\") " pod="openshift-console/console-f9d7485db-f8l78" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.722670 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/317b7923-187f-47a4-9b82-53b20c86d9a4-service-ca\") pod \"console-f9d7485db-f8l78\" (UID: \"317b7923-187f-47a4-9b82-53b20c86d9a4\") " pod="openshift-console/console-f9d7485db-f8l78" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.722689 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/317b7923-187f-47a4-9b82-53b20c86d9a4-trusted-ca-bundle\") pod \"console-f9d7485db-f8l78\" (UID: \"317b7923-187f-47a4-9b82-53b20c86d9a4\") " pod="openshift-console/console-f9d7485db-f8l78" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.722718 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/74e9c10d-c66e-48ea-9813-cb221012dda4-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-pj8qs\" (UID: \"74e9c10d-c66e-48ea-9813-cb221012dda4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pj8qs" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.722756 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/4849186f-21ae-4654-a813-4b176093abbd-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-rdfcg\" (UID: \"4849186f-21ae-4654-a813-4b176093abbd\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rdfcg" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.722778 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f7829943-cabc-49c5-a547-62124e38d49e-client-ca\") pod \"route-controller-manager-6576b87f9c-98h6l\" (UID: \"f7829943-cabc-49c5-a547-62124e38d49e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-98h6l" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.722821 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2b42593e-4396-4b1a-9272-ca2a2fb5ad6e-encryption-config\") pod \"apiserver-7bbb656c7d-l5vld\" (UID: \"2b42593e-4396-4b1a-9272-ca2a2fb5ad6e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l5vld" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.722863 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-t4x4n" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.723088 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-mq75m" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.723760 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/317b7923-187f-47a4-9b82-53b20c86d9a4-service-ca\") pod \"console-f9d7485db-f8l78\" (UID: \"317b7923-187f-47a4-9b82-53b20c86d9a4\") " pod="openshift-console/console-f9d7485db-f8l78" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.724273 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/317b7923-187f-47a4-9b82-53b20c86d9a4-trusted-ca-bundle\") pod \"console-f9d7485db-f8l78\" (UID: \"317b7923-187f-47a4-9b82-53b20c86d9a4\") " pod="openshift-console/console-f9d7485db-f8l78" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.722865 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hphpt\" (UniqueName: \"kubernetes.io/projected/a00e530f-3c79-40da-a3f3-faf504af3fea-kube-api-access-hphpt\") pod \"openshift-config-operator-7777fb866f-gm548\" (UID: \"a00e530f-3c79-40da-a3f3-faf504af3fea\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gm548" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.724604 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/317b7923-187f-47a4-9b82-53b20c86d9a4-console-serving-cert\") pod \"console-f9d7485db-f8l78\" (UID: \"317b7923-187f-47a4-9b82-53b20c86d9a4\") " pod="openshift-console/console-f9d7485db-f8l78" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.724640 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/317b7923-187f-47a4-9b82-53b20c86d9a4-oauth-serving-cert\") pod \"console-f9d7485db-f8l78\" (UID: \"317b7923-187f-47a4-9b82-53b20c86d9a4\") " pod="openshift-console/console-f9d7485db-f8l78" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.724665 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/56fd5dd1-12e2-4b22-a4ab-291d6bb3ce92-bound-sa-token\") pod \"ingress-operator-5b745b69d9-r76vv\" (UID: \"56fd5dd1-12e2-4b22-a4ab-291d6bb3ce92\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-r76vv" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.724686 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/317b7923-187f-47a4-9b82-53b20c86d9a4-console-config\") pod \"console-f9d7485db-f8l78\" (UID: \"317b7923-187f-47a4-9b82-53b20c86d9a4\") " pod="openshift-console/console-f9d7485db-f8l78" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.724730 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/745ab1fe-8d37-420f-9c12-62cd7a33aeaa-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-hkkwq\" (UID: \"745ab1fe-8d37-420f-9c12-62cd7a33aeaa\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hkkwq" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.724754 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgzxc\" (UniqueName: \"kubernetes.io/projected/4849186f-21ae-4654-a813-4b176093abbd-kube-api-access-pgzxc\") pod \"cluster-samples-operator-665b6dd947-rdfcg\" (UID: \"4849186f-21ae-4654-a813-4b176093abbd\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rdfcg" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.724808 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2b42593e-4396-4b1a-9272-ca2a2fb5ad6e-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-l5vld\" (UID: \"2b42593e-4396-4b1a-9272-ca2a2fb5ad6e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l5vld" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.724836 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b42593e-4396-4b1a-9272-ca2a2fb5ad6e-serving-cert\") pod \"apiserver-7bbb656c7d-l5vld\" (UID: \"2b42593e-4396-4b1a-9272-ca2a2fb5ad6e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l5vld" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.724862 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/56fd5dd1-12e2-4b22-a4ab-291d6bb3ce92-metrics-tls\") pod \"ingress-operator-5b745b69d9-r76vv\" (UID: \"56fd5dd1-12e2-4b22-a4ab-291d6bb3ce92\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-r76vv" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.724893 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/745ab1fe-8d37-420f-9c12-62cd7a33aeaa-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-hkkwq\" (UID: \"745ab1fe-8d37-420f-9c12-62cd7a33aeaa\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hkkwq" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.724913 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2b42593e-4396-4b1a-9272-ca2a2fb5ad6e-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-l5vld\" (UID: \"2b42593e-4396-4b1a-9272-ca2a2fb5ad6e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l5vld" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.724934 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqq2z\" (UniqueName: \"kubernetes.io/projected/bd697754-4524-44ed-be19-18747c0626f5-kube-api-access-qqq2z\") pod \"control-plane-machine-set-operator-78cbb6b69f-dkktg\" (UID: \"bd697754-4524-44ed-be19-18747c0626f5\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dkktg" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.724957 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c2b2606-ba50-42a8-ae0b-2176b853cb94-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-tdwfd\" (UID: \"3c2b2606-ba50-42a8-ae0b-2176b853cb94\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tdwfd" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.724978 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggbks\" (UniqueName: \"kubernetes.io/projected/3c2b2606-ba50-42a8-ae0b-2176b853cb94-kube-api-access-ggbks\") pod \"openshift-controller-manager-operator-756b6f6bc6-tdwfd\" (UID: \"3c2b2606-ba50-42a8-ae0b-2176b853cb94\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tdwfd" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.724999 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/74e9c10d-c66e-48ea-9813-cb221012dda4-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-pj8qs\" (UID: \"74e9c10d-c66e-48ea-9813-cb221012dda4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pj8qs" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.725017 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4e2257ac-9d7a-4e81-b560-32f28a73830e-metrics-tls\") pod \"dns-operator-744455d44c-b68qv\" (UID: \"4e2257ac-9d7a-4e81-b560-32f28a73830e\") " pod="openshift-dns-operator/dns-operator-744455d44c-b68qv" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.725036 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2b42593e-4396-4b1a-9272-ca2a2fb5ad6e-audit-dir\") pod \"apiserver-7bbb656c7d-l5vld\" (UID: \"2b42593e-4396-4b1a-9272-ca2a2fb5ad6e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l5vld" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.725074 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7829943-cabc-49c5-a547-62124e38d49e-config\") pod \"route-controller-manager-6576b87f9c-98h6l\" (UID: \"f7829943-cabc-49c5-a547-62124e38d49e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-98h6l" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.725096 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74e9c10d-c66e-48ea-9813-cb221012dda4-config\") pod \"kube-apiserver-operator-766d6c64bb-pj8qs\" (UID: \"74e9c10d-c66e-48ea-9813-cb221012dda4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pj8qs" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.725125 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a00e530f-3c79-40da-a3f3-faf504af3fea-serving-cert\") pod \"openshift-config-operator-7777fb866f-gm548\" (UID: \"a00e530f-3c79-40da-a3f3-faf504af3fea\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gm548" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.725157 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3c2b2606-ba50-42a8-ae0b-2176b853cb94-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-tdwfd\" (UID: \"3c2b2606-ba50-42a8-ae0b-2176b853cb94\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tdwfd" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.725198 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7829943-cabc-49c5-a547-62124e38d49e-serving-cert\") pod \"route-controller-manager-6576b87f9c-98h6l\" (UID: \"f7829943-cabc-49c5-a547-62124e38d49e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-98h6l" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.725222 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/a00e530f-3c79-40da-a3f3-faf504af3fea-available-featuregates\") pod \"openshift-config-operator-7777fb866f-gm548\" (UID: \"a00e530f-3c79-40da-a3f3-faf504af3fea\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gm548" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.725440 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/745ab1fe-8d37-420f-9c12-62cd7a33aeaa-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-hkkwq\" (UID: \"745ab1fe-8d37-420f-9c12-62cd7a33aeaa\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hkkwq" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.725563 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xhg5n" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.725610 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/a00e530f-3c79-40da-a3f3-faf504af3fea-available-featuregates\") pod \"openshift-config-operator-7777fb866f-gm548\" (UID: \"a00e530f-3c79-40da-a3f3-faf504af3fea\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gm548" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.726144 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/4849186f-21ae-4654-a813-4b176093abbd-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-rdfcg\" (UID: \"4849186f-21ae-4654-a813-4b176093abbd\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rdfcg" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.726596 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c2b2606-ba50-42a8-ae0b-2176b853cb94-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-tdwfd\" (UID: \"3c2b2606-ba50-42a8-ae0b-2176b853cb94\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tdwfd" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.726823 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2b42593e-4396-4b1a-9272-ca2a2fb5ad6e-audit-dir\") pod \"apiserver-7bbb656c7d-l5vld\" (UID: \"2b42593e-4396-4b1a-9272-ca2a2fb5ad6e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l5vld" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.726841 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/317b7923-187f-47a4-9b82-53b20c86d9a4-console-config\") pod \"console-f9d7485db-f8l78\" (UID: \"317b7923-187f-47a4-9b82-53b20c86d9a4\") " pod="openshift-console/console-f9d7485db-f8l78" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.727563 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2b42593e-4396-4b1a-9272-ca2a2fb5ad6e-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-l5vld\" (UID: \"2b42593e-4396-4b1a-9272-ca2a2fb5ad6e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l5vld" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.727617 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2b42593e-4396-4b1a-9272-ca2a2fb5ad6e-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-l5vld\" (UID: \"2b42593e-4396-4b1a-9272-ca2a2fb5ad6e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l5vld" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.728185 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7829943-cabc-49c5-a547-62124e38d49e-config\") pod \"route-controller-manager-6576b87f9c-98h6l\" (UID: \"f7829943-cabc-49c5-a547-62124e38d49e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-98h6l" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.728305 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/317b7923-187f-47a4-9b82-53b20c86d9a4-console-oauth-config\") pod \"console-f9d7485db-f8l78\" (UID: \"317b7923-187f-47a4-9b82-53b20c86d9a4\") " pod="openshift-console/console-f9d7485db-f8l78" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.728845 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2b42593e-4396-4b1a-9272-ca2a2fb5ad6e-etcd-client\") pod \"apiserver-7bbb656c7d-l5vld\" (UID: \"2b42593e-4396-4b1a-9272-ca2a2fb5ad6e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l5vld" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.729135 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a16e3671-5260-4e88-8fcc-8ba2252af6ea-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-mt8x9\" (UID: \"a16e3671-5260-4e88-8fcc-8ba2252af6ea\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mt8x9" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.729299 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.729478 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f7829943-cabc-49c5-a547-62124e38d49e-client-ca\") pod \"route-controller-manager-6576b87f9c-98h6l\" (UID: \"f7829943-cabc-49c5-a547-62124e38d49e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-98h6l" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.729493 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7718b7b0-f1a7-4b87-8ebe-345e9100463d-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-m62b2\" (UID: \"7718b7b0-f1a7-4b87-8ebe-345e9100463d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-m62b2" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.730030 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/317b7923-187f-47a4-9b82-53b20c86d9a4-oauth-serving-cert\") pod \"console-f9d7485db-f8l78\" (UID: \"317b7923-187f-47a4-9b82-53b20c86d9a4\") " pod="openshift-console/console-f9d7485db-f8l78" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.730250 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.731044 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wpqs8"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.731589 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wpqs8" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.732629 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.732909 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-l2gg8"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.733476 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.733769 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/56fd5dd1-12e2-4b22-a4ab-291d6bb3ce92-metrics-tls\") pod \"ingress-operator-5b745b69d9-r76vv\" (UID: \"56fd5dd1-12e2-4b22-a4ab-291d6bb3ce92\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-r76vv" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.733839 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2b42593e-4396-4b1a-9272-ca2a2fb5ad6e-encryption-config\") pod \"apiserver-7bbb656c7d-l5vld\" (UID: \"2b42593e-4396-4b1a-9272-ca2a2fb5ad6e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l5vld" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.734695 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b9dce599-89b8-4c6c-a12c-65f9ef64de9b-trusted-ca-bundle\") pod \"apiserver-76f77b778f-wd9qj\" (UID: \"b9dce599-89b8-4c6c-a12c-65f9ef64de9b\") " pod="openshift-apiserver/apiserver-76f77b778f-wd9qj" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.734936 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a00e530f-3c79-40da-a3f3-faf504af3fea-serving-cert\") pod \"openshift-config-operator-7777fb866f-gm548\" (UID: \"a00e530f-3c79-40da-a3f3-faf504af3fea\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gm548" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.735728 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3c2b2606-ba50-42a8-ae0b-2176b853cb94-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-tdwfd\" (UID: \"3c2b2606-ba50-42a8-ae0b-2176b853cb94\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tdwfd" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.735744 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vbpch"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.735904 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-l2gg8" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.736025 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-wvgmn"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.736206 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vbpch" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.736257 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4e2257ac-9d7a-4e81-b560-32f28a73830e-metrics-tls\") pod \"dns-operator-744455d44c-b68qv\" (UID: \"4e2257ac-9d7a-4e81-b560-32f28a73830e\") " pod="openshift-dns-operator/dns-operator-744455d44c-b68qv" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.736359 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-wvgmn" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.737409 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500755-fjlmd"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.737656 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b42593e-4396-4b1a-9272-ca2a2fb5ad6e-serving-cert\") pod \"apiserver-7bbb656c7d-l5vld\" (UID: \"2b42593e-4396-4b1a-9272-ca2a2fb5ad6e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l5vld" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.737871 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500755-fjlmd" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.738235 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vgtrr"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.738589 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vgtrr" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.738991 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-qt46z"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.739479 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7829943-cabc-49c5-a547-62124e38d49e-serving-cert\") pod \"route-controller-manager-6576b87f9c-98h6l\" (UID: \"f7829943-cabc-49c5-a547-62124e38d49e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-98h6l" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.740131 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hfp26"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.740183 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-qt46z" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.741904 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-4cq4j\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " pod="openshift-authentication/oauth-openshift-558db77b4-4cq4j" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.744853 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-csdnv"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.746361 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hfp26" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.748459 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/317b7923-187f-47a4-9b82-53b20c86d9a4-console-serving-cert\") pod \"console-f9d7485db-f8l78\" (UID: \"317b7923-187f-47a4-9b82-53b20c86d9a4\") " pod="openshift-console/console-f9d7485db-f8l78" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.748474 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/745ab1fe-8d37-420f-9c12-62cd7a33aeaa-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-hkkwq\" (UID: \"745ab1fe-8d37-420f-9c12-62cd7a33aeaa\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hkkwq" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.750201 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kxfm5"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.750303 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-csdnv" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.753561 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.754144 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-c6qfs"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.754894 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r8cnf"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.754918 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-l595n"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.756210 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-l595n" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.756552 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kxfm5" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.756852 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-c6qfs" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.758626 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.760326 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-4cq4j\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " pod="openshift-authentication/oauth-openshift-558db77b4-4cq4j" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.760476 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-gmgdk"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.761478 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-xk69x"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.762432 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.762733 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-gmgdk" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.772258 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-74k4s"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.772642 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xk69x" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.773402 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/56fd5dd1-12e2-4b22-a4ab-291d6bb3ce92-trusted-ca\") pod \"ingress-operator-5b745b69d9-r76vv\" (UID: \"56fd5dd1-12e2-4b22-a4ab-291d6bb3ce92\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-r76vv" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.775740 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-4cq4j"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.777478 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-4bdpn"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.778094 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-4bdpn" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.781071 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.783936 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-b68qv"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.785356 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-l5vld"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.790055 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-mt8x9"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.794459 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-98h6l"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.795711 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-ct5l6"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.796662 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-t4x4n"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.798143 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-l2gg8"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.799223 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.799352 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tdwfd"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.800302 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-gm548"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.801997 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-wd9qj"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.802510 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-m62b2"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.803559 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dkktg"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.804549 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pmbhw"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.805506 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hkkwq"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.807350 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xhg5n"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.808748 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-f8l78"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.811528 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vbpch"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.812672 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rdfcg"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.813899 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-mw8sl"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.815015 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-mw8sl" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.815275 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500755-fjlmd"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.816640 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-hpgjp"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.818014 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vgtrr"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.818093 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.818888 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-gmgdk"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.820143 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-wvgmn"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.821592 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-c6qfs"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.823067 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wpqs8"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.823988 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hfp26"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.825242 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-qt46z"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.825655 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a8928569-d787-4825-883d-6adf585f069b-profile-collector-cert\") pod \"olm-operator-6b444d44fb-wpqs8\" (UID: \"a8928569-d787-4825-883d-6adf585f069b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wpqs8" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.825873 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/74e9c10d-c66e-48ea-9813-cb221012dda4-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-pj8qs\" (UID: \"74e9c10d-c66e-48ea-9813-cb221012dda4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pj8qs" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.826768 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kxfm5"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.826968 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twbv5\" (UniqueName: \"kubernetes.io/projected/613678ea-f8f5-4fc1-beb2-0fde3dfe77ac-kube-api-access-twbv5\") pod \"multus-admission-controller-857f4d67dd-l595n\" (UID: \"613678ea-f8f5-4fc1-beb2-0fde3dfe77ac\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-l595n" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.827148 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/f5ff4cdb-523e-49f2-9d0f-54eca688b223-csi-data-dir\") pod \"csi-hostpathplugin-mw8sl\" (UID: \"f5ff4cdb-523e-49f2-9d0f-54eca688b223\") " pod="hostpath-provisioner/csi-hostpathplugin-mw8sl" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.827240 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/c7442bb0-8bba-418d-a9e1-7f935fb07257-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-xhg5n\" (UID: \"c7442bb0-8bba-418d-a9e1-7f935fb07257\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xhg5n" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.827324 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/613678ea-f8f5-4fc1-beb2-0fde3dfe77ac-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-l595n\" (UID: \"613678ea-f8f5-4fc1-beb2-0fde3dfe77ac\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-l595n" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.827420 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dec3c184-4636-43b5-9cc4-2188635c2719-config\") pod \"etcd-operator-b45778765-t4x4n\" (UID: \"dec3c184-4636-43b5-9cc4-2188635c2719\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t4x4n" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.827512 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/dec3c184-4636-43b5-9cc4-2188635c2719-etcd-client\") pod \"etcd-operator-b45778765-t4x4n\" (UID: \"dec3c184-4636-43b5-9cc4-2188635c2719\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t4x4n" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.827591 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a11da899-1807-4360-be80-ea30f6ef2c5f-srv-cert\") pod \"catalog-operator-68c6474976-vbpch\" (UID: \"a11da899-1807-4360-be80-ea30f6ef2c5f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vbpch" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.827666 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjcqw\" (UniqueName: \"kubernetes.io/projected/a0d837f7-b35f-4879-b0f9-d82411bf16ff-kube-api-access-cjcqw\") pod \"packageserver-d55dfcdfc-kxfm5\" (UID: \"a0d837f7-b35f-4879-b0f9-d82411bf16ff\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kxfm5" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.827730 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxzfn\" (UniqueName: \"kubernetes.io/projected/e886b855-272d-435e-ae5b-1498f637409d-kube-api-access-qxzfn\") pod \"dns-default-4bdpn\" (UID: \"e886b855-272d-435e-ae5b-1498f637409d\") " pod="openshift-dns/dns-default-4bdpn" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.827696 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pj8qs"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.827857 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e18d552a-2c5e-4671-a594-dd443cf9eacc-images\") pod \"machine-config-operator-74547568cd-xk69x\" (UID: \"e18d552a-2c5e-4671-a594-dd443cf9eacc\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xk69x" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.827929 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e886b855-272d-435e-ae5b-1498f637409d-metrics-tls\") pod \"dns-default-4bdpn\" (UID: \"e886b855-272d-435e-ae5b-1498f637409d\") " pod="openshift-dns/dns-default-4bdpn" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.827999 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e18d552a-2c5e-4671-a594-dd443cf9eacc-proxy-tls\") pod \"machine-config-operator-74547568cd-xk69x\" (UID: \"e18d552a-2c5e-4671-a594-dd443cf9eacc\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xk69x" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.828068 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a0d837f7-b35f-4879-b0f9-d82411bf16ff-apiservice-cert\") pod \"packageserver-d55dfcdfc-kxfm5\" (UID: \"a0d837f7-b35f-4879-b0f9-d82411bf16ff\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kxfm5" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.828132 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkl7b\" (UniqueName: \"kubernetes.io/projected/2c71221f-f379-4474-b1b5-b948c663a1d1-kube-api-access-qkl7b\") pod \"machine-config-controller-84d6567774-l2gg8\" (UID: \"2c71221f-f379-4474-b1b5-b948c663a1d1\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-l2gg8" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.828260 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dec3c184-4636-43b5-9cc4-2188635c2719-serving-cert\") pod \"etcd-operator-b45778765-t4x4n\" (UID: \"dec3c184-4636-43b5-9cc4-2188635c2719\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t4x4n" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.828359 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a8928569-d787-4825-883d-6adf585f069b-srv-cert\") pod \"olm-operator-6b444d44fb-wpqs8\" (UID: \"a8928569-d787-4825-883d-6adf585f069b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wpqs8" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.828433 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfl9v\" (UniqueName: \"kubernetes.io/projected/c7442bb0-8bba-418d-a9e1-7f935fb07257-kube-api-access-dfl9v\") pod \"package-server-manager-789f6589d5-xhg5n\" (UID: \"c7442bb0-8bba-418d-a9e1-7f935fb07257\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xhg5n" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.828504 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2c71221f-f379-4474-b1b5-b948c663a1d1-proxy-tls\") pod \"machine-config-controller-84d6567774-l2gg8\" (UID: \"2c71221f-f379-4474-b1b5-b948c663a1d1\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-l2gg8" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.828585 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/f5ff4cdb-523e-49f2-9d0f-54eca688b223-plugins-dir\") pod \"csi-hostpathplugin-mw8sl\" (UID: \"f5ff4cdb-523e-49f2-9d0f-54eca688b223\") " pod="hostpath-provisioner/csi-hostpathplugin-mw8sl" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.828689 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e18d552a-2c5e-4671-a594-dd443cf9eacc-auth-proxy-config\") pod \"machine-config-operator-74547568cd-xk69x\" (UID: \"e18d552a-2c5e-4671-a594-dd443cf9eacc\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xk69x" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.828722 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-knzns\" (UniqueName: \"kubernetes.io/projected/f5ff4cdb-523e-49f2-9d0f-54eca688b223-kube-api-access-knzns\") pod \"csi-hostpathplugin-mw8sl\" (UID: \"f5ff4cdb-523e-49f2-9d0f-54eca688b223\") " pod="hostpath-provisioner/csi-hostpathplugin-mw8sl" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.828760 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dgd6\" (UniqueName: \"kubernetes.io/projected/a11da899-1807-4360-be80-ea30f6ef2c5f-kube-api-access-7dgd6\") pod \"catalog-operator-68c6474976-vbpch\" (UID: \"a11da899-1807-4360-be80-ea30f6ef2c5f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vbpch" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.828791 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/b2deaba5-c009-4df6-87b9-185b9a8addc5-signing-cabundle\") pod \"service-ca-9c57cc56f-c6qfs\" (UID: \"b2deaba5-c009-4df6-87b9-185b9a8addc5\") " pod="openshift-service-ca/service-ca-9c57cc56f-c6qfs" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.828813 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpb9v\" (UniqueName: \"kubernetes.io/projected/b2deaba5-c009-4df6-87b9-185b9a8addc5-kube-api-access-lpb9v\") pod \"service-ca-9c57cc56f-c6qfs\" (UID: \"b2deaba5-c009-4df6-87b9-185b9a8addc5\") " pod="openshift-service-ca/service-ca-9c57cc56f-c6qfs" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.828834 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/f5ff4cdb-523e-49f2-9d0f-54eca688b223-mountpoint-dir\") pod \"csi-hostpathplugin-mw8sl\" (UID: \"f5ff4cdb-523e-49f2-9d0f-54eca688b223\") " pod="hostpath-provisioner/csi-hostpathplugin-mw8sl" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.828888 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e886b855-272d-435e-ae5b-1498f637409d-config-volume\") pod \"dns-default-4bdpn\" (UID: \"e886b855-272d-435e-ae5b-1498f637409d\") " pod="openshift-dns/dns-default-4bdpn" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.828934 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/f5ff4cdb-523e-49f2-9d0f-54eca688b223-socket-dir\") pod \"csi-hostpathplugin-mw8sl\" (UID: \"f5ff4cdb-523e-49f2-9d0f-54eca688b223\") " pod="hostpath-provisioner/csi-hostpathplugin-mw8sl" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.829496 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/f5ff4cdb-523e-49f2-9d0f-54eca688b223-registration-dir\") pod \"csi-hostpathplugin-mw8sl\" (UID: \"f5ff4cdb-523e-49f2-9d0f-54eca688b223\") " pod="hostpath-provisioner/csi-hostpathplugin-mw8sl" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.829538 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9j9k\" (UniqueName: \"kubernetes.io/projected/a8928569-d787-4825-883d-6adf585f069b-kube-api-access-h9j9k\") pod \"olm-operator-6b444d44fb-wpqs8\" (UID: \"a8928569-d787-4825-883d-6adf585f069b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wpqs8" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.829332 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-xk69x"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.829606 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/b2deaba5-c009-4df6-87b9-185b9a8addc5-signing-key\") pod \"service-ca-9c57cc56f-c6qfs\" (UID: \"b2deaba5-c009-4df6-87b9-185b9a8addc5\") " pod="openshift-service-ca/service-ca-9c57cc56f-c6qfs" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.829719 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a0d837f7-b35f-4879-b0f9-d82411bf16ff-tmpfs\") pod \"packageserver-d55dfcdfc-kxfm5\" (UID: \"a0d837f7-b35f-4879-b0f9-d82411bf16ff\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kxfm5" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.829763 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/dec3c184-4636-43b5-9cc4-2188635c2719-etcd-service-ca\") pod \"etcd-operator-b45778765-t4x4n\" (UID: \"dec3c184-4636-43b5-9cc4-2188635c2719\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t4x4n" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.829812 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlcd5\" (UniqueName: \"kubernetes.io/projected/dec3c184-4636-43b5-9cc4-2188635c2719-kube-api-access-hlcd5\") pod \"etcd-operator-b45778765-t4x4n\" (UID: \"dec3c184-4636-43b5-9cc4-2188635c2719\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t4x4n" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.829889 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2c71221f-f379-4474-b1b5-b948c663a1d1-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-l2gg8\" (UID: \"2c71221f-f379-4474-b1b5-b948c663a1d1\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-l2gg8" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.829921 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6cv9\" (UniqueName: \"kubernetes.io/projected/e18d552a-2c5e-4671-a594-dd443cf9eacc-kube-api-access-z6cv9\") pod \"machine-config-operator-74547568cd-xk69x\" (UID: \"e18d552a-2c5e-4671-a594-dd443cf9eacc\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xk69x" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.830060 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-csdnv\" (UID: \"ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e\") " pod="openshift-marketplace/marketplace-operator-79b997595-csdnv" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.830197 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a0d837f7-b35f-4879-b0f9-d82411bf16ff-webhook-cert\") pod \"packageserver-d55dfcdfc-kxfm5\" (UID: \"a0d837f7-b35f-4879-b0f9-d82411bf16ff\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kxfm5" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.830224 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-csdnv\" (UID: \"ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e\") " pod="openshift-marketplace/marketplace-operator-79b997595-csdnv" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.830277 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/dec3c184-4636-43b5-9cc4-2188635c2719-etcd-ca\") pod \"etcd-operator-b45778765-t4x4n\" (UID: \"dec3c184-4636-43b5-9cc4-2188635c2719\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t4x4n" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.830305 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a11da899-1807-4360-be80-ea30f6ef2c5f-profile-collector-cert\") pod \"catalog-operator-68c6474976-vbpch\" (UID: \"a11da899-1807-4360-be80-ea30f6ef2c5f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vbpch" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.830384 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lvtx\" (UniqueName: \"kubernetes.io/projected/ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e-kube-api-access-7lvtx\") pod \"marketplace-operator-79b997595-csdnv\" (UID: \"ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e\") " pod="openshift-marketplace/marketplace-operator-79b997595-csdnv" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.830990 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-csdnv"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.832219 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-r76vv"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.833436 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-mw8sl"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.834640 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-4bdpn"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.836023 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-l595n"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.837250 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-nzjsl"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.838348 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-nzjsl" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.838846 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.840277 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-nzjsl"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.841298 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-j6ht9"] Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.841790 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-j6ht9" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.848194 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74e9c10d-c66e-48ea-9813-cb221012dda4-config\") pod \"kube-apiserver-operator-766d6c64bb-pj8qs\" (UID: \"74e9c10d-c66e-48ea-9813-cb221012dda4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pj8qs" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.858298 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.869609 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/bd697754-4524-44ed-be19-18747c0626f5-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-dkktg\" (UID: \"bd697754-4524-44ed-be19-18747c0626f5\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dkktg" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.878437 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.911157 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7bvj\" (UniqueName: \"kubernetes.io/projected/dab28707-1fc8-45a4-8ad4-0842b7f70239-kube-api-access-t7bvj\") pod \"console-operator-58897d9998-ct5l6\" (UID: \"dab28707-1fc8-45a4-8ad4-0842b7f70239\") " pod="openshift-console-operator/console-operator-58897d9998-ct5l6" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.931183 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/b2deaba5-c009-4df6-87b9-185b9a8addc5-signing-key\") pod \"service-ca-9c57cc56f-c6qfs\" (UID: \"b2deaba5-c009-4df6-87b9-185b9a8addc5\") " pod="openshift-service-ca/service-ca-9c57cc56f-c6qfs" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.931426 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a0d837f7-b35f-4879-b0f9-d82411bf16ff-tmpfs\") pod \"packageserver-d55dfcdfc-kxfm5\" (UID: \"a0d837f7-b35f-4879-b0f9-d82411bf16ff\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kxfm5" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.931599 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/dec3c184-4636-43b5-9cc4-2188635c2719-etcd-service-ca\") pod \"etcd-operator-b45778765-t4x4n\" (UID: \"dec3c184-4636-43b5-9cc4-2188635c2719\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t4x4n" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.931721 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlcd5\" (UniqueName: \"kubernetes.io/projected/dec3c184-4636-43b5-9cc4-2188635c2719-kube-api-access-hlcd5\") pod \"etcd-operator-b45778765-t4x4n\" (UID: \"dec3c184-4636-43b5-9cc4-2188635c2719\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t4x4n" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.931841 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2c71221f-f379-4474-b1b5-b948c663a1d1-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-l2gg8\" (UID: \"2c71221f-f379-4474-b1b5-b948c663a1d1\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-l2gg8" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.931964 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6cv9\" (UniqueName: \"kubernetes.io/projected/e18d552a-2c5e-4671-a594-dd443cf9eacc-kube-api-access-z6cv9\") pod \"machine-config-operator-74547568cd-xk69x\" (UID: \"e18d552a-2c5e-4671-a594-dd443cf9eacc\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xk69x" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.932092 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-csdnv\" (UID: \"ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e\") " pod="openshift-marketplace/marketplace-operator-79b997595-csdnv" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.932300 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a0d837f7-b35f-4879-b0f9-d82411bf16ff-webhook-cert\") pod \"packageserver-d55dfcdfc-kxfm5\" (UID: \"a0d837f7-b35f-4879-b0f9-d82411bf16ff\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kxfm5" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.932426 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-csdnv\" (UID: \"ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e\") " pod="openshift-marketplace/marketplace-operator-79b997595-csdnv" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.932539 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/dec3c184-4636-43b5-9cc4-2188635c2719-etcd-ca\") pod \"etcd-operator-b45778765-t4x4n\" (UID: \"dec3c184-4636-43b5-9cc4-2188635c2719\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t4x4n" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.932355 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a0d837f7-b35f-4879-b0f9-d82411bf16ff-tmpfs\") pod \"packageserver-d55dfcdfc-kxfm5\" (UID: \"a0d837f7-b35f-4879-b0f9-d82411bf16ff\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kxfm5" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.932655 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a11da899-1807-4360-be80-ea30f6ef2c5f-profile-collector-cert\") pod \"catalog-operator-68c6474976-vbpch\" (UID: \"a11da899-1807-4360-be80-ea30f6ef2c5f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vbpch" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.932762 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lvtx\" (UniqueName: \"kubernetes.io/projected/ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e-kube-api-access-7lvtx\") pod \"marketplace-operator-79b997595-csdnv\" (UID: \"ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e\") " pod="openshift-marketplace/marketplace-operator-79b997595-csdnv" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.932808 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a8928569-d787-4825-883d-6adf585f069b-profile-collector-cert\") pod \"olm-operator-6b444d44fb-wpqs8\" (UID: \"a8928569-d787-4825-883d-6adf585f069b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wpqs8" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.932824 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9w6sz\" (UniqueName: \"kubernetes.io/projected/a3e78085-1f5d-4f0c-84bb-62b787b1fbdd-kube-api-access-9w6sz\") pod \"machine-api-operator-5694c8668f-74k4s\" (UID: \"a3e78085-1f5d-4f0c-84bb-62b787b1fbdd\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-74k4s" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.932843 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twbv5\" (UniqueName: \"kubernetes.io/projected/613678ea-f8f5-4fc1-beb2-0fde3dfe77ac-kube-api-access-twbv5\") pod \"multus-admission-controller-857f4d67dd-l595n\" (UID: \"613678ea-f8f5-4fc1-beb2-0fde3dfe77ac\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-l595n" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.932876 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/f5ff4cdb-523e-49f2-9d0f-54eca688b223-csi-data-dir\") pod \"csi-hostpathplugin-mw8sl\" (UID: \"f5ff4cdb-523e-49f2-9d0f-54eca688b223\") " pod="hostpath-provisioner/csi-hostpathplugin-mw8sl" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.932914 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/c7442bb0-8bba-418d-a9e1-7f935fb07257-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-xhg5n\" (UID: \"c7442bb0-8bba-418d-a9e1-7f935fb07257\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xhg5n" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.932934 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/f5ff4cdb-523e-49f2-9d0f-54eca688b223-csi-data-dir\") pod \"csi-hostpathplugin-mw8sl\" (UID: \"f5ff4cdb-523e-49f2-9d0f-54eca688b223\") " pod="hostpath-provisioner/csi-hostpathplugin-mw8sl" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.932948 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/613678ea-f8f5-4fc1-beb2-0fde3dfe77ac-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-l595n\" (UID: \"613678ea-f8f5-4fc1-beb2-0fde3dfe77ac\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-l595n" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.932984 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dec3c184-4636-43b5-9cc4-2188635c2719-config\") pod \"etcd-operator-b45778765-t4x4n\" (UID: \"dec3c184-4636-43b5-9cc4-2188635c2719\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t4x4n" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.933013 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/dec3c184-4636-43b5-9cc4-2188635c2719-etcd-client\") pod \"etcd-operator-b45778765-t4x4n\" (UID: \"dec3c184-4636-43b5-9cc4-2188635c2719\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t4x4n" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.933045 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a11da899-1807-4360-be80-ea30f6ef2c5f-srv-cert\") pod \"catalog-operator-68c6474976-vbpch\" (UID: \"a11da899-1807-4360-be80-ea30f6ef2c5f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vbpch" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.933079 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjcqw\" (UniqueName: \"kubernetes.io/projected/a0d837f7-b35f-4879-b0f9-d82411bf16ff-kube-api-access-cjcqw\") pod \"packageserver-d55dfcdfc-kxfm5\" (UID: \"a0d837f7-b35f-4879-b0f9-d82411bf16ff\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kxfm5" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.933112 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxzfn\" (UniqueName: \"kubernetes.io/projected/e886b855-272d-435e-ae5b-1498f637409d-kube-api-access-qxzfn\") pod \"dns-default-4bdpn\" (UID: \"e886b855-272d-435e-ae5b-1498f637409d\") " pod="openshift-dns/dns-default-4bdpn" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.933147 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e18d552a-2c5e-4671-a594-dd443cf9eacc-images\") pod \"machine-config-operator-74547568cd-xk69x\" (UID: \"e18d552a-2c5e-4671-a594-dd443cf9eacc\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xk69x" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.933201 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e18d552a-2c5e-4671-a594-dd443cf9eacc-proxy-tls\") pod \"machine-config-operator-74547568cd-xk69x\" (UID: \"e18d552a-2c5e-4671-a594-dd443cf9eacc\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xk69x" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.933231 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e886b855-272d-435e-ae5b-1498f637409d-metrics-tls\") pod \"dns-default-4bdpn\" (UID: \"e886b855-272d-435e-ae5b-1498f637409d\") " pod="openshift-dns/dns-default-4bdpn" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.933292 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a0d837f7-b35f-4879-b0f9-d82411bf16ff-apiservice-cert\") pod \"packageserver-d55dfcdfc-kxfm5\" (UID: \"a0d837f7-b35f-4879-b0f9-d82411bf16ff\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kxfm5" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.933316 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2c71221f-f379-4474-b1b5-b948c663a1d1-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-l2gg8\" (UID: \"2c71221f-f379-4474-b1b5-b948c663a1d1\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-l2gg8" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.933332 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qkl7b\" (UniqueName: \"kubernetes.io/projected/2c71221f-f379-4474-b1b5-b948c663a1d1-kube-api-access-qkl7b\") pod \"machine-config-controller-84d6567774-l2gg8\" (UID: \"2c71221f-f379-4474-b1b5-b948c663a1d1\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-l2gg8" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.933408 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dec3c184-4636-43b5-9cc4-2188635c2719-serving-cert\") pod \"etcd-operator-b45778765-t4x4n\" (UID: \"dec3c184-4636-43b5-9cc4-2188635c2719\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t4x4n" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.933474 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a8928569-d787-4825-883d-6adf585f069b-srv-cert\") pod \"olm-operator-6b444d44fb-wpqs8\" (UID: \"a8928569-d787-4825-883d-6adf585f069b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wpqs8" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.933523 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfl9v\" (UniqueName: \"kubernetes.io/projected/c7442bb0-8bba-418d-a9e1-7f935fb07257-kube-api-access-dfl9v\") pod \"package-server-manager-789f6589d5-xhg5n\" (UID: \"c7442bb0-8bba-418d-a9e1-7f935fb07257\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xhg5n" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.933554 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2c71221f-f379-4474-b1b5-b948c663a1d1-proxy-tls\") pod \"machine-config-controller-84d6567774-l2gg8\" (UID: \"2c71221f-f379-4474-b1b5-b948c663a1d1\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-l2gg8" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.933585 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/f5ff4cdb-523e-49f2-9d0f-54eca688b223-plugins-dir\") pod \"csi-hostpathplugin-mw8sl\" (UID: \"f5ff4cdb-523e-49f2-9d0f-54eca688b223\") " pod="hostpath-provisioner/csi-hostpathplugin-mw8sl" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.933628 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e18d552a-2c5e-4671-a594-dd443cf9eacc-auth-proxy-config\") pod \"machine-config-operator-74547568cd-xk69x\" (UID: \"e18d552a-2c5e-4671-a594-dd443cf9eacc\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xk69x" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.933658 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-knzns\" (UniqueName: \"kubernetes.io/projected/f5ff4cdb-523e-49f2-9d0f-54eca688b223-kube-api-access-knzns\") pod \"csi-hostpathplugin-mw8sl\" (UID: \"f5ff4cdb-523e-49f2-9d0f-54eca688b223\") " pod="hostpath-provisioner/csi-hostpathplugin-mw8sl" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.933694 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dgd6\" (UniqueName: \"kubernetes.io/projected/a11da899-1807-4360-be80-ea30f6ef2c5f-kube-api-access-7dgd6\") pod \"catalog-operator-68c6474976-vbpch\" (UID: \"a11da899-1807-4360-be80-ea30f6ef2c5f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vbpch" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.933725 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/b2deaba5-c009-4df6-87b9-185b9a8addc5-signing-cabundle\") pod \"service-ca-9c57cc56f-c6qfs\" (UID: \"b2deaba5-c009-4df6-87b9-185b9a8addc5\") " pod="openshift-service-ca/service-ca-9c57cc56f-c6qfs" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.933756 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpb9v\" (UniqueName: \"kubernetes.io/projected/b2deaba5-c009-4df6-87b9-185b9a8addc5-kube-api-access-lpb9v\") pod \"service-ca-9c57cc56f-c6qfs\" (UID: \"b2deaba5-c009-4df6-87b9-185b9a8addc5\") " pod="openshift-service-ca/service-ca-9c57cc56f-c6qfs" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.933789 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/f5ff4cdb-523e-49f2-9d0f-54eca688b223-mountpoint-dir\") pod \"csi-hostpathplugin-mw8sl\" (UID: \"f5ff4cdb-523e-49f2-9d0f-54eca688b223\") " pod="hostpath-provisioner/csi-hostpathplugin-mw8sl" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.933832 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e886b855-272d-435e-ae5b-1498f637409d-config-volume\") pod \"dns-default-4bdpn\" (UID: \"e886b855-272d-435e-ae5b-1498f637409d\") " pod="openshift-dns/dns-default-4bdpn" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.933882 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/f5ff4cdb-523e-49f2-9d0f-54eca688b223-socket-dir\") pod \"csi-hostpathplugin-mw8sl\" (UID: \"f5ff4cdb-523e-49f2-9d0f-54eca688b223\") " pod="hostpath-provisioner/csi-hostpathplugin-mw8sl" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.933911 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/f5ff4cdb-523e-49f2-9d0f-54eca688b223-registration-dir\") pod \"csi-hostpathplugin-mw8sl\" (UID: \"f5ff4cdb-523e-49f2-9d0f-54eca688b223\") " pod="hostpath-provisioner/csi-hostpathplugin-mw8sl" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.933953 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9j9k\" (UniqueName: \"kubernetes.io/projected/a8928569-d787-4825-883d-6adf585f069b-kube-api-access-h9j9k\") pod \"olm-operator-6b444d44fb-wpqs8\" (UID: \"a8928569-d787-4825-883d-6adf585f069b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wpqs8" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.934136 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/f5ff4cdb-523e-49f2-9d0f-54eca688b223-plugins-dir\") pod \"csi-hostpathplugin-mw8sl\" (UID: \"f5ff4cdb-523e-49f2-9d0f-54eca688b223\") " pod="hostpath-provisioner/csi-hostpathplugin-mw8sl" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.934257 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/f5ff4cdb-523e-49f2-9d0f-54eca688b223-mountpoint-dir\") pod \"csi-hostpathplugin-mw8sl\" (UID: \"f5ff4cdb-523e-49f2-9d0f-54eca688b223\") " pod="hostpath-provisioner/csi-hostpathplugin-mw8sl" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.934269 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/f5ff4cdb-523e-49f2-9d0f-54eca688b223-socket-dir\") pod \"csi-hostpathplugin-mw8sl\" (UID: \"f5ff4cdb-523e-49f2-9d0f-54eca688b223\") " pod="hostpath-provisioner/csi-hostpathplugin-mw8sl" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.934304 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/f5ff4cdb-523e-49f2-9d0f-54eca688b223-registration-dir\") pod \"csi-hostpathplugin-mw8sl\" (UID: \"f5ff4cdb-523e-49f2-9d0f-54eca688b223\") " pod="hostpath-provisioner/csi-hostpathplugin-mw8sl" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.934843 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e18d552a-2c5e-4671-a594-dd443cf9eacc-auth-proxy-config\") pod \"machine-config-operator-74547568cd-xk69x\" (UID: \"e18d552a-2c5e-4671-a594-dd443cf9eacc\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xk69x" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.951892 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfvl2\" (UniqueName: \"kubernetes.io/projected/7718b7b0-f1a7-4b87-8ebe-345e9100463d-kube-api-access-vfvl2\") pod \"controller-manager-879f6c89f-m62b2\" (UID: \"7718b7b0-f1a7-4b87-8ebe-345e9100463d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-m62b2" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.985760 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbxp6\" (UniqueName: \"kubernetes.io/projected/d8fddc32-a0a6-4149-aa39-b7f51f78a772-kube-api-access-rbxp6\") pod \"machine-approver-56656f9798-79hfl\" (UID: \"d8fddc32-a0a6-4149-aa39-b7f51f78a772\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-79hfl" Feb 02 15:17:39 crc kubenswrapper[4733]: I0202 15:17:39.995590 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8c5w\" (UniqueName: \"kubernetes.io/projected/a16e3671-5260-4e88-8fcc-8ba2252af6ea-kube-api-access-g8c5w\") pod \"authentication-operator-69f744f599-mt8x9\" (UID: \"a16e3671-5260-4e88-8fcc-8ba2252af6ea\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mt8x9" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.016230 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ztcgq\" (UniqueName: \"kubernetes.io/projected/d667dd66-2215-4d9c-9cf5-64ea2f9a2dc8-kube-api-access-ztcgq\") pod \"openshift-apiserver-operator-796bbdcf4f-r8cnf\" (UID: \"d667dd66-2215-4d9c-9cf5-64ea2f9a2dc8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r8cnf" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.034753 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pppg7\" (UniqueName: \"kubernetes.io/projected/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-kube-api-access-pppg7\") pod \"oauth-openshift-558db77b4-4cq4j\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " pod="openshift-authentication/oauth-openshift-558db77b4-4cq4j" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.048487 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-4cq4j" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.055113 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzwg2\" (UniqueName: \"kubernetes.io/projected/b9dce599-89b8-4c6c-a12c-65f9ef64de9b-kube-api-access-lzwg2\") pod \"apiserver-76f77b778f-wd9qj\" (UID: \"b9dce599-89b8-4c6c-a12c-65f9ef64de9b\") " pod="openshift-apiserver/apiserver-76f77b778f-wd9qj" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.063887 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-ct5l6" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.077283 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-m62b2" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.082647 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-wd9qj" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.100496 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5f8gt\" (UniqueName: \"kubernetes.io/projected/f7829943-cabc-49c5-a547-62124e38d49e-kube-api-access-5f8gt\") pod \"route-controller-manager-6576b87f9c-98h6l\" (UID: \"f7829943-cabc-49c5-a547-62124e38d49e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-98h6l" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.103923 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-79hfl" Feb 02 15:17:40 crc kubenswrapper[4733]: W0202 15:17:40.125861 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd8fddc32_a0a6_4149_aa39_b7f51f78a772.slice/crio-f194d12b66ec31047a8512be0d908dbdccd5cc353c19ac489c25781727028ece WatchSource:0}: Error finding container f194d12b66ec31047a8512be0d908dbdccd5cc353c19ac489c25781727028ece: Status 404 returned error can't find the container with id f194d12b66ec31047a8512be0d908dbdccd5cc353c19ac489c25781727028ece Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.136401 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-74k4s" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.136431 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xclz\" (UniqueName: \"kubernetes.io/projected/4e2257ac-9d7a-4e81-b560-32f28a73830e-kube-api-access-4xclz\") pod \"dns-operator-744455d44c-b68qv\" (UID: \"4e2257ac-9d7a-4e81-b560-32f28a73830e\") " pod="openshift-dns-operator/dns-operator-744455d44c-b68qv" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.139002 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.143512 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r8cnf" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.150678 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dw6wr\" (UniqueName: \"kubernetes.io/projected/317b7923-187f-47a4-9b82-53b20c86d9a4-kube-api-access-dw6wr\") pod \"console-f9d7485db-f8l78\" (UID: \"317b7923-187f-47a4-9b82-53b20c86d9a4\") " pod="openshift-console/console-f9d7485db-f8l78" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.162808 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.194278 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-mt8x9" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.217314 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdbpq\" (UniqueName: \"kubernetes.io/projected/745ab1fe-8d37-420f-9c12-62cd7a33aeaa-kube-api-access-mdbpq\") pod \"cluster-image-registry-operator-dc59b4c8b-hkkwq\" (UID: \"745ab1fe-8d37-420f-9c12-62cd7a33aeaa\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hkkwq" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.219866 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.222596 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6rpm\" (UniqueName: \"kubernetes.io/projected/2b42593e-4396-4b1a-9272-ca2a2fb5ad6e-kube-api-access-k6rpm\") pod \"apiserver-7bbb656c7d-l5vld\" (UID: \"2b42593e-4396-4b1a-9272-ca2a2fb5ad6e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l5vld" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.238778 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.281475 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.281875 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tw8r\" (UniqueName: \"kubernetes.io/projected/56fd5dd1-12e2-4b22-a4ab-291d6bb3ce92-kube-api-access-2tw8r\") pod \"ingress-operator-5b745b69d9-r76vv\" (UID: \"56fd5dd1-12e2-4b22-a4ab-291d6bb3ce92\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-r76vv" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.299473 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.318597 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.355537 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hphpt\" (UniqueName: \"kubernetes.io/projected/a00e530f-3c79-40da-a3f3-faf504af3fea-kube-api-access-hphpt\") pod \"openshift-config-operator-7777fb866f-gm548\" (UID: \"a00e530f-3c79-40da-a3f3-faf504af3fea\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gm548" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.359317 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.378707 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.383951 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/dec3c184-4636-43b5-9cc4-2188635c2719-etcd-ca\") pod \"etcd-operator-b45778765-t4x4n\" (UID: \"dec3c184-4636-43b5-9cc4-2188635c2719\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t4x4n" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.389860 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-b68qv" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.396283 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-98h6l" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.399513 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.405145 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l5vld" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.412721 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-f8l78" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.419715 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.427690 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dec3c184-4636-43b5-9cc4-2188635c2719-serving-cert\") pod \"etcd-operator-b45778765-t4x4n\" (UID: \"dec3c184-4636-43b5-9cc4-2188635c2719\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t4x4n" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.440355 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.447968 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/dec3c184-4636-43b5-9cc4-2188635c2719-etcd-client\") pod \"etcd-operator-b45778765-t4x4n\" (UID: \"dec3c184-4636-43b5-9cc4-2188635c2719\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t4x4n" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.459150 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.464434 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dec3c184-4636-43b5-9cc4-2188635c2719-config\") pod \"etcd-operator-b45778765-t4x4n\" (UID: \"dec3c184-4636-43b5-9cc4-2188635c2719\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t4x4n" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.498296 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqq2z\" (UniqueName: \"kubernetes.io/projected/bd697754-4524-44ed-be19-18747c0626f5-kube-api-access-qqq2z\") pod \"control-plane-machine-set-operator-78cbb6b69f-dkktg\" (UID: \"bd697754-4524-44ed-be19-18747c0626f5\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dkktg" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.498758 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.502678 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/dec3c184-4636-43b5-9cc4-2188635c2719-etcd-service-ca\") pod \"etcd-operator-b45778765-t4x4n\" (UID: \"dec3c184-4636-43b5-9cc4-2188635c2719\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t4x4n" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.520069 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.528525 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gm548" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.546344 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-4cq4j"] Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.548157 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-m62b2"] Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.552892 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/56fd5dd1-12e2-4b22-a4ab-291d6bb3ce92-bound-sa-token\") pod \"ingress-operator-5b745b69d9-r76vv\" (UID: \"56fd5dd1-12e2-4b22-a4ab-291d6bb3ce92\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-r76vv" Feb 02 15:17:40 crc kubenswrapper[4733]: W0202 15:17:40.568400 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7718b7b0_f1a7_4b87_8ebe_345e9100463d.slice/crio-52c3bfd7078308d02908ed10f8d54d9f00bd3465c4fc389e2c149c066cbec839 WatchSource:0}: Error finding container 52c3bfd7078308d02908ed10f8d54d9f00bd3465c4fc389e2c149c066cbec839: Status 404 returned error can't find the container with id 52c3bfd7078308d02908ed10f8d54d9f00bd3465c4fc389e2c149c066cbec839 Feb 02 15:17:40 crc kubenswrapper[4733]: W0202 15:17:40.568750 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7a9e1fe8_5666_4dae_baa9_df9fee6e21f1.slice/crio-953ec4cf5eb753fc046706b992d748a9b4c63d57a2b621951ceff34fcbac7021 WatchSource:0}: Error finding container 953ec4cf5eb753fc046706b992d748a9b4c63d57a2b621951ceff34fcbac7021: Status 404 returned error can't find the container with id 953ec4cf5eb753fc046706b992d748a9b4c63d57a2b621951ceff34fcbac7021 Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.579532 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggbks\" (UniqueName: \"kubernetes.io/projected/3c2b2606-ba50-42a8-ae0b-2176b853cb94-kube-api-access-ggbks\") pod \"openshift-controller-manager-operator-756b6f6bc6-tdwfd\" (UID: \"3c2b2606-ba50-42a8-ae0b-2176b853cb94\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tdwfd" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.600195 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/74e9c10d-c66e-48ea-9813-cb221012dda4-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-pj8qs\" (UID: \"74e9c10d-c66e-48ea-9813-cb221012dda4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pj8qs" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.602436 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.616087 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-ct5l6"] Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.617912 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-wd9qj"] Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.634656 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgzxc\" (UniqueName: \"kubernetes.io/projected/4849186f-21ae-4654-a813-4b176093abbd-kube-api-access-pgzxc\") pod \"cluster-samples-operator-665b6dd947-rdfcg\" (UID: \"4849186f-21ae-4654-a813-4b176093abbd\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rdfcg" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.638726 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rdfcg" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.664783 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.677021 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r8cnf"] Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.678536 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.680975 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/745ab1fe-8d37-420f-9c12-62cd7a33aeaa-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-hkkwq\" (UID: \"745ab1fe-8d37-420f-9c12-62cd7a33aeaa\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hkkwq" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.690186 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-74k4s"] Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.691501 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-mt8x9"] Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.698588 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.719003 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hkkwq" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.720114 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.729562 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tdwfd" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.735572 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-r76vv" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.736918 4733 request.go:700] Waited for 1.005712675s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ingress/configmaps?fieldSelector=metadata.name%3Dservice-ca-bundle&limit=500&resourceVersion=0 Feb 02 15:17:40 crc kubenswrapper[4733]: W0202 15:17:40.737559 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd667dd66_2215_4d9c_9cf5_64ea2f9a2dc8.slice/crio-d244bda36ab53689df07b48c479f2aaf51f4ae98200f269dddbfa72af50629fc WatchSource:0}: Error finding container d244bda36ab53689df07b48c479f2aaf51f4ae98200f269dddbfa72af50629fc: Status 404 returned error can't find the container with id d244bda36ab53689df07b48c479f2aaf51f4ae98200f269dddbfa72af50629fc Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.737856 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-l5vld"] Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.738685 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.747734 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pj8qs" Feb 02 15:17:40 crc kubenswrapper[4733]: W0202 15:17:40.750754 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda3e78085_1f5d_4f0c_84bb_62b787b1fbdd.slice/crio-060b867460d82d53df4e6339d2ceac7920daa5e8e6802a16386f4be67b0eab91 WatchSource:0}: Error finding container 060b867460d82d53df4e6339d2ceac7920daa5e8e6802a16386f4be67b0eab91: Status 404 returned error can't find the container with id 060b867460d82d53df4e6339d2ceac7920daa5e8e6802a16386f4be67b0eab91 Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.760611 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.763134 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dkktg" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.779292 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Feb 02 15:17:40 crc kubenswrapper[4733]: W0202 15:17:40.786947 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2b42593e_4396_4b1a_9272_ca2a2fb5ad6e.slice/crio-b9c99a993be7a07102d8fdfa10d9c0039c59c14805f9680f335f06aa69b5ebd0 WatchSource:0}: Error finding container b9c99a993be7a07102d8fdfa10d9c0039c59c14805f9680f335f06aa69b5ebd0: Status 404 returned error can't find the container with id b9c99a993be7a07102d8fdfa10d9c0039c59c14805f9680f335f06aa69b5ebd0 Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.800805 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.801098 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-gm548"] Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.807438 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/c7442bb0-8bba-418d-a9e1-7f935fb07257-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-xhg5n\" (UID: \"c7442bb0-8bba-418d-a9e1-7f935fb07257\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xhg5n" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.819879 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.840983 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.845173 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-f8l78"] Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.867259 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-b68qv"] Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.871107 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.876234 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a11da899-1807-4360-be80-ea30f6ef2c5f-profile-collector-cert\") pod \"catalog-operator-68c6474976-vbpch\" (UID: \"a11da899-1807-4360-be80-ea30f6ef2c5f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vbpch" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.876352 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a8928569-d787-4825-883d-6adf585f069b-profile-collector-cert\") pod \"olm-operator-6b444d44fb-wpqs8\" (UID: \"a8928569-d787-4825-883d-6adf585f069b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wpqs8" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.882544 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.883539 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-98h6l"] Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.892570 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rdfcg"] Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.893655 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a8928569-d787-4825-883d-6adf585f069b-srv-cert\") pod \"olm-operator-6b444d44fb-wpqs8\" (UID: \"a8928569-d787-4825-883d-6adf585f069b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wpqs8" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.901420 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.911049 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2c71221f-f379-4474-b1b5-b948c663a1d1-proxy-tls\") pod \"machine-config-controller-84d6567774-l2gg8\" (UID: \"2c71221f-f379-4474-b1b5-b948c663a1d1\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-l2gg8" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.919707 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Feb 02 15:17:40 crc kubenswrapper[4733]: E0202 15:17:40.931877 4733 secret.go:188] Couldn't get secret openshift-service-ca/signing-key: failed to sync secret cache: timed out waiting for the condition Feb 02 15:17:40 crc kubenswrapper[4733]: E0202 15:17:40.931958 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b2deaba5-c009-4df6-87b9-185b9a8addc5-signing-key podName:b2deaba5-c009-4df6-87b9-185b9a8addc5 nodeName:}" failed. No retries permitted until 2026-02-02 15:17:41.431937441 +0000 UTC m=+144.883398799 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "signing-key" (UniqueName: "kubernetes.io/secret/b2deaba5-c009-4df6-87b9-185b9a8addc5-signing-key") pod "service-ca-9c57cc56f-c6qfs" (UID: "b2deaba5-c009-4df6-87b9-185b9a8addc5") : failed to sync secret cache: timed out waiting for the condition Feb 02 15:17:40 crc kubenswrapper[4733]: E0202 15:17:40.932248 4733 secret.go:188] Couldn't get secret openshift-marketplace/marketplace-operator-metrics: failed to sync secret cache: timed out waiting for the condition Feb 02 15:17:40 crc kubenswrapper[4733]: E0202 15:17:40.932321 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e-marketplace-operator-metrics podName:ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e nodeName:}" failed. No retries permitted until 2026-02-02 15:17:41.432303751 +0000 UTC m=+144.883765109 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "marketplace-operator-metrics" (UniqueName: "kubernetes.io/secret/ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e-marketplace-operator-metrics") pod "marketplace-operator-79b997595-csdnv" (UID: "ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e") : failed to sync secret cache: timed out waiting for the condition Feb 02 15:17:40 crc kubenswrapper[4733]: E0202 15:17:40.932952 4733 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/packageserver-service-cert: failed to sync secret cache: timed out waiting for the condition Feb 02 15:17:40 crc kubenswrapper[4733]: E0202 15:17:40.932988 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0d837f7-b35f-4879-b0f9-d82411bf16ff-webhook-cert podName:a0d837f7-b35f-4879-b0f9-d82411bf16ff nodeName:}" failed. No retries permitted until 2026-02-02 15:17:41.432979879 +0000 UTC m=+144.884441227 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-cert" (UniqueName: "kubernetes.io/secret/a0d837f7-b35f-4879-b0f9-d82411bf16ff-webhook-cert") pod "packageserver-d55dfcdfc-kxfm5" (UID: "a0d837f7-b35f-4879-b0f9-d82411bf16ff") : failed to sync secret cache: timed out waiting for the condition Feb 02 15:17:40 crc kubenswrapper[4733]: E0202 15:17:40.932993 4733 configmap.go:193] Couldn't get configMap openshift-marketplace/marketplace-trusted-ca: failed to sync configmap cache: timed out waiting for the condition Feb 02 15:17:40 crc kubenswrapper[4733]: E0202 15:17:40.933082 4733 secret.go:188] Couldn't get secret openshift-multus/multus-admission-controller-secret: failed to sync secret cache: timed out waiting for the condition Feb 02 15:17:40 crc kubenswrapper[4733]: E0202 15:17:40.933085 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e-marketplace-trusted-ca podName:ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e nodeName:}" failed. No retries permitted until 2026-02-02 15:17:41.433063362 +0000 UTC m=+144.884524820 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "marketplace-trusted-ca" (UniqueName: "kubernetes.io/configmap/ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e-marketplace-trusted-ca") pod "marketplace-operator-79b997595-csdnv" (UID: "ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e") : failed to sync configmap cache: timed out waiting for the condition Feb 02 15:17:40 crc kubenswrapper[4733]: E0202 15:17:40.933111 4733 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/catalog-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Feb 02 15:17:40 crc kubenswrapper[4733]: E0202 15:17:40.933132 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/613678ea-f8f5-4fc1-beb2-0fde3dfe77ac-webhook-certs podName:613678ea-f8f5-4fc1-beb2-0fde3dfe77ac nodeName:}" failed. No retries permitted until 2026-02-02 15:17:41.433122733 +0000 UTC m=+144.884584091 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/613678ea-f8f5-4fc1-beb2-0fde3dfe77ac-webhook-certs") pod "multus-admission-controller-857f4d67dd-l595n" (UID: "613678ea-f8f5-4fc1-beb2-0fde3dfe77ac") : failed to sync secret cache: timed out waiting for the condition Feb 02 15:17:40 crc kubenswrapper[4733]: E0202 15:17:40.933148 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a11da899-1807-4360-be80-ea30f6ef2c5f-srv-cert podName:a11da899-1807-4360-be80-ea30f6ef2c5f nodeName:}" failed. No retries permitted until 2026-02-02 15:17:41.433139254 +0000 UTC m=+144.884600612 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "srv-cert" (UniqueName: "kubernetes.io/secret/a11da899-1807-4360-be80-ea30f6ef2c5f-srv-cert") pod "catalog-operator-68c6474976-vbpch" (UID: "a11da899-1807-4360-be80-ea30f6ef2c5f") : failed to sync secret cache: timed out waiting for the condition Feb 02 15:17:40 crc kubenswrapper[4733]: E0202 15:17:40.934081 4733 secret.go:188] Couldn't get secret openshift-dns/dns-default-metrics-tls: failed to sync secret cache: timed out waiting for the condition Feb 02 15:17:40 crc kubenswrapper[4733]: E0202 15:17:40.934121 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e886b855-272d-435e-ae5b-1498f637409d-metrics-tls podName:e886b855-272d-435e-ae5b-1498f637409d nodeName:}" failed. No retries permitted until 2026-02-02 15:17:41.434102729 +0000 UTC m=+144.885564077 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/e886b855-272d-435e-ae5b-1498f637409d-metrics-tls") pod "dns-default-4bdpn" (UID: "e886b855-272d-435e-ae5b-1498f637409d") : failed to sync secret cache: timed out waiting for the condition Feb 02 15:17:40 crc kubenswrapper[4733]: E0202 15:17:40.934139 4733 secret.go:188] Couldn't get secret openshift-machine-config-operator/mco-proxy-tls: failed to sync secret cache: timed out waiting for the condition Feb 02 15:17:40 crc kubenswrapper[4733]: E0202 15:17:40.934174 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e18d552a-2c5e-4671-a594-dd443cf9eacc-proxy-tls podName:e18d552a-2c5e-4671-a594-dd443cf9eacc nodeName:}" failed. No retries permitted until 2026-02-02 15:17:41.434154511 +0000 UTC m=+144.885615869 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "proxy-tls" (UniqueName: "kubernetes.io/secret/e18d552a-2c5e-4671-a594-dd443cf9eacc-proxy-tls") pod "machine-config-operator-74547568cd-xk69x" (UID: "e18d552a-2c5e-4671-a594-dd443cf9eacc") : failed to sync secret cache: timed out waiting for the condition Feb 02 15:17:40 crc kubenswrapper[4733]: E0202 15:17:40.934200 4733 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/packageserver-service-cert: failed to sync secret cache: timed out waiting for the condition Feb 02 15:17:40 crc kubenswrapper[4733]: E0202 15:17:40.934221 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0d837f7-b35f-4879-b0f9-d82411bf16ff-apiservice-cert podName:a0d837f7-b35f-4879-b0f9-d82411bf16ff nodeName:}" failed. No retries permitted until 2026-02-02 15:17:41.434214642 +0000 UTC m=+144.885676000 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "apiservice-cert" (UniqueName: "kubernetes.io/secret/a0d837f7-b35f-4879-b0f9-d82411bf16ff-apiservice-cert") pod "packageserver-d55dfcdfc-kxfm5" (UID: "a0d837f7-b35f-4879-b0f9-d82411bf16ff") : failed to sync secret cache: timed out waiting for the condition Feb 02 15:17:40 crc kubenswrapper[4733]: E0202 15:17:40.934262 4733 configmap.go:193] Couldn't get configMap openshift-machine-config-operator/machine-config-operator-images: failed to sync configmap cache: timed out waiting for the condition Feb 02 15:17:40 crc kubenswrapper[4733]: E0202 15:17:40.934280 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/e18d552a-2c5e-4671-a594-dd443cf9eacc-images podName:e18d552a-2c5e-4671-a594-dd443cf9eacc nodeName:}" failed. No retries permitted until 2026-02-02 15:17:41.434275034 +0000 UTC m=+144.885736382 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "images" (UniqueName: "kubernetes.io/configmap/e18d552a-2c5e-4671-a594-dd443cf9eacc-images") pod "machine-config-operator-74547568cd-xk69x" (UID: "e18d552a-2c5e-4671-a594-dd443cf9eacc") : failed to sync configmap cache: timed out waiting for the condition Feb 02 15:17:40 crc kubenswrapper[4733]: E0202 15:17:40.934302 4733 configmap.go:193] Couldn't get configMap openshift-dns/dns-default: failed to sync configmap cache: timed out waiting for the condition Feb 02 15:17:40 crc kubenswrapper[4733]: E0202 15:17:40.934319 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/e886b855-272d-435e-ae5b-1498f637409d-config-volume podName:e886b855-272d-435e-ae5b-1498f637409d nodeName:}" failed. No retries permitted until 2026-02-02 15:17:41.434313915 +0000 UTC m=+144.885775273 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/e886b855-272d-435e-ae5b-1498f637409d-config-volume") pod "dns-default-4bdpn" (UID: "e886b855-272d-435e-ae5b-1498f637409d") : failed to sync configmap cache: timed out waiting for the condition Feb 02 15:17:40 crc kubenswrapper[4733]: E0202 15:17:40.934342 4733 configmap.go:193] Couldn't get configMap openshift-service-ca/signing-cabundle: failed to sync configmap cache: timed out waiting for the condition Feb 02 15:17:40 crc kubenswrapper[4733]: E0202 15:17:40.934377 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b2deaba5-c009-4df6-87b9-185b9a8addc5-signing-cabundle podName:b2deaba5-c009-4df6-87b9-185b9a8addc5 nodeName:}" failed. No retries permitted until 2026-02-02 15:17:41.434354116 +0000 UTC m=+144.885815474 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "signing-cabundle" (UniqueName: "kubernetes.io/configmap/b2deaba5-c009-4df6-87b9-185b9a8addc5-signing-cabundle") pod "service-ca-9c57cc56f-c6qfs" (UID: "b2deaba5-c009-4df6-87b9-185b9a8addc5") : failed to sync configmap cache: timed out waiting for the condition Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.940298 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.958564 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.978771 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 02 15:17:40 crc kubenswrapper[4733]: I0202 15:17:40.999424 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.018124 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.038977 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.045206 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-4cq4j" event={"ID":"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1","Type":"ContainerStarted","Data":"953ec4cf5eb753fc046706b992d748a9b4c63d57a2b621951ceff34fcbac7021"} Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.046770 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-f8l78" event={"ID":"317b7923-187f-47a4-9b82-53b20c86d9a4","Type":"ContainerStarted","Data":"3db713ca9ab1c5e1a1a2012194a340cceee30884619c798315a0f1153f30e9a3"} Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.048176 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gm548" event={"ID":"a00e530f-3c79-40da-a3f3-faf504af3fea","Type":"ContainerStarted","Data":"ba1a7591e97e38110e28216e9bf4c185c895734552c8ee2098334cdfcf083328"} Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.049813 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-79hfl" event={"ID":"d8fddc32-a0a6-4149-aa39-b7f51f78a772","Type":"ContainerStarted","Data":"6f0c4ebce1225166f1b527b79f75d5d89329856e87e76b1b74bb21e7a1fcefa0"} Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.049839 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-79hfl" event={"ID":"d8fddc32-a0a6-4149-aa39-b7f51f78a772","Type":"ContainerStarted","Data":"72a26054dc0b9ee5da6388b80d5e4acead919465849557545913bd52c33379cf"} Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.049849 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-79hfl" event={"ID":"d8fddc32-a0a6-4149-aa39-b7f51f78a772","Type":"ContainerStarted","Data":"f194d12b66ec31047a8512be0d908dbdccd5cc353c19ac489c25781727028ece"} Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.050734 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-74k4s" event={"ID":"a3e78085-1f5d-4f0c-84bb-62b787b1fbdd","Type":"ContainerStarted","Data":"060b867460d82d53df4e6339d2ceac7920daa5e8e6802a16386f4be67b0eab91"} Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.059874 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.064860 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-b68qv" event={"ID":"4e2257ac-9d7a-4e81-b560-32f28a73830e","Type":"ContainerStarted","Data":"b8be51da9b30992cc23cd56d6bad8d7b363faf603afea07537b2067fea950523"} Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.068793 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r8cnf" event={"ID":"d667dd66-2215-4d9c-9cf5-64ea2f9a2dc8","Type":"ContainerStarted","Data":"d244bda36ab53689df07b48c479f2aaf51f4ae98200f269dddbfa72af50629fc"} Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.069300 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-r76vv"] Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.080648 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.084012 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-ct5l6" event={"ID":"dab28707-1fc8-45a4-8ad4-0842b7f70239","Type":"ContainerStarted","Data":"af4f0bc6047119e122148cd9f72c217e0cefd8c0aa61d4e6b62a6a9e7894aae2"} Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.084049 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-ct5l6" event={"ID":"dab28707-1fc8-45a4-8ad4-0842b7f70239","Type":"ContainerStarted","Data":"ec25830f9cdf0b43b87decd58a0004ac103137fd028b1cb27b14b8901f3b271d"} Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.084983 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-ct5l6" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.085990 4733 patch_prober.go:28] interesting pod/console-operator-58897d9998-ct5l6 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/readyz\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.086038 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-ct5l6" podUID="dab28707-1fc8-45a4-8ad4-0842b7f70239" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.9:8443/readyz\": dial tcp 10.217.0.9:8443: connect: connection refused" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.099116 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l5vld" event={"ID":"2b42593e-4396-4b1a-9272-ca2a2fb5ad6e","Type":"ContainerStarted","Data":"b9c99a993be7a07102d8fdfa10d9c0039c59c14805f9680f335f06aa69b5ebd0"} Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.099271 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.101545 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-98h6l" event={"ID":"f7829943-cabc-49c5-a547-62124e38d49e","Type":"ContainerStarted","Data":"69b3eabefd5f9ed969de63b29afef1d1720ae7a64d3b673e908abe27a6e2fc02"} Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.102824 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-wd9qj" event={"ID":"b9dce599-89b8-4c6c-a12c-65f9ef64de9b","Type":"ContainerStarted","Data":"dcffd0ed367c6227e3a43cd98ca806af80acef37a55082ce41a5b5a4e1a7e1f7"} Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.105876 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-m62b2" event={"ID":"7718b7b0-f1a7-4b87-8ebe-345e9100463d","Type":"ContainerStarted","Data":"e1d6f6e0975a0529556931f0c74afc5d70e93d838245956f2238a399298a6400"} Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.105933 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-m62b2" event={"ID":"7718b7b0-f1a7-4b87-8ebe-345e9100463d","Type":"ContainerStarted","Data":"52c3bfd7078308d02908ed10f8d54d9f00bd3465c4fc389e2c149c066cbec839"} Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.106801 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-m62b2" Feb 02 15:17:41 crc kubenswrapper[4733]: W0202 15:17:41.108513 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod56fd5dd1_12e2_4b22_a4ab_291d6bb3ce92.slice/crio-4a4ea4ca8820d4f471e8c1ec6e26d5d417b4627afa4d727141cd3bc33dcb283d WatchSource:0}: Error finding container 4a4ea4ca8820d4f471e8c1ec6e26d5d417b4627afa4d727141cd3bc33dcb283d: Status 404 returned error can't find the container with id 4a4ea4ca8820d4f471e8c1ec6e26d5d417b4627afa4d727141cd3bc33dcb283d Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.108592 4733 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-m62b2 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.20:8443/healthz\": dial tcp 10.217.0.20:8443: connect: connection refused" start-of-body= Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.108635 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-m62b2" podUID="7718b7b0-f1a7-4b87-8ebe-345e9100463d" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.20:8443/healthz\": dial tcp 10.217.0.20:8443: connect: connection refused" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.109537 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-mt8x9" event={"ID":"a16e3671-5260-4e88-8fcc-8ba2252af6ea","Type":"ContainerStarted","Data":"0481ac3e30b916b50dc04f461990f20fa63ee42e79d7289761696b402ecd076b"} Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.131335 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.139231 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.159990 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.179657 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.198673 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.222003 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.226153 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pj8qs"] Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.239835 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.249866 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hkkwq"] Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.258247 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.278063 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.299566 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.319425 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.333131 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tdwfd"] Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.337473 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dkktg"] Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.351551 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.360314 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.378671 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.401875 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.419465 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.439005 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.459442 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.478866 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.479487 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e886b855-272d-435e-ae5b-1498f637409d-config-volume\") pod \"dns-default-4bdpn\" (UID: \"e886b855-272d-435e-ae5b-1498f637409d\") " pod="openshift-dns/dns-default-4bdpn" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.479545 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/b2deaba5-c009-4df6-87b9-185b9a8addc5-signing-key\") pod \"service-ca-9c57cc56f-c6qfs\" (UID: \"b2deaba5-c009-4df6-87b9-185b9a8addc5\") " pod="openshift-service-ca/service-ca-9c57cc56f-c6qfs" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.479624 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-csdnv\" (UID: \"ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e\") " pod="openshift-marketplace/marketplace-operator-79b997595-csdnv" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.479676 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a0d837f7-b35f-4879-b0f9-d82411bf16ff-webhook-cert\") pod \"packageserver-d55dfcdfc-kxfm5\" (UID: \"a0d837f7-b35f-4879-b0f9-d82411bf16ff\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kxfm5" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.479695 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-csdnv\" (UID: \"ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e\") " pod="openshift-marketplace/marketplace-operator-79b997595-csdnv" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.479768 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/613678ea-f8f5-4fc1-beb2-0fde3dfe77ac-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-l595n\" (UID: \"613678ea-f8f5-4fc1-beb2-0fde3dfe77ac\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-l595n" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.479792 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a11da899-1807-4360-be80-ea30f6ef2c5f-srv-cert\") pod \"catalog-operator-68c6474976-vbpch\" (UID: \"a11da899-1807-4360-be80-ea30f6ef2c5f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vbpch" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.479821 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e18d552a-2c5e-4671-a594-dd443cf9eacc-images\") pod \"machine-config-operator-74547568cd-xk69x\" (UID: \"e18d552a-2c5e-4671-a594-dd443cf9eacc\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xk69x" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.479838 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e18d552a-2c5e-4671-a594-dd443cf9eacc-proxy-tls\") pod \"machine-config-operator-74547568cd-xk69x\" (UID: \"e18d552a-2c5e-4671-a594-dd443cf9eacc\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xk69x" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.479856 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e886b855-272d-435e-ae5b-1498f637409d-metrics-tls\") pod \"dns-default-4bdpn\" (UID: \"e886b855-272d-435e-ae5b-1498f637409d\") " pod="openshift-dns/dns-default-4bdpn" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.479872 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a0d837f7-b35f-4879-b0f9-d82411bf16ff-apiservice-cert\") pod \"packageserver-d55dfcdfc-kxfm5\" (UID: \"a0d837f7-b35f-4879-b0f9-d82411bf16ff\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kxfm5" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.479927 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/b2deaba5-c009-4df6-87b9-185b9a8addc5-signing-cabundle\") pod \"service-ca-9c57cc56f-c6qfs\" (UID: \"b2deaba5-c009-4df6-87b9-185b9a8addc5\") " pod="openshift-service-ca/service-ca-9c57cc56f-c6qfs" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.481711 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/b2deaba5-c009-4df6-87b9-185b9a8addc5-signing-cabundle\") pod \"service-ca-9c57cc56f-c6qfs\" (UID: \"b2deaba5-c009-4df6-87b9-185b9a8addc5\") " pod="openshift-service-ca/service-ca-9c57cc56f-c6qfs" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.482353 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-csdnv\" (UID: \"ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e\") " pod="openshift-marketplace/marketplace-operator-79b997595-csdnv" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.486720 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-csdnv\" (UID: \"ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e\") " pod="openshift-marketplace/marketplace-operator-79b997595-csdnv" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.486766 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a0d837f7-b35f-4879-b0f9-d82411bf16ff-apiservice-cert\") pod \"packageserver-d55dfcdfc-kxfm5\" (UID: \"a0d837f7-b35f-4879-b0f9-d82411bf16ff\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kxfm5" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.486923 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a0d837f7-b35f-4879-b0f9-d82411bf16ff-webhook-cert\") pod \"packageserver-d55dfcdfc-kxfm5\" (UID: \"a0d837f7-b35f-4879-b0f9-d82411bf16ff\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kxfm5" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.486981 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a11da899-1807-4360-be80-ea30f6ef2c5f-srv-cert\") pod \"catalog-operator-68c6474976-vbpch\" (UID: \"a11da899-1807-4360-be80-ea30f6ef2c5f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vbpch" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.487058 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/b2deaba5-c009-4df6-87b9-185b9a8addc5-signing-key\") pod \"service-ca-9c57cc56f-c6qfs\" (UID: \"b2deaba5-c009-4df6-87b9-185b9a8addc5\") " pod="openshift-service-ca/service-ca-9c57cc56f-c6qfs" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.499047 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.499426 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/613678ea-f8f5-4fc1-beb2-0fde3dfe77ac-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-l595n\" (UID: \"613678ea-f8f5-4fc1-beb2-0fde3dfe77ac\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-l595n" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.506551 4733 csr.go:261] certificate signing request csr-fvqwb is approved, waiting to be issued Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.514031 4733 csr.go:257] certificate signing request csr-fvqwb is issued Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.521311 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.538833 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.559697 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.579198 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.598608 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.619607 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.621246 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e18d552a-2c5e-4671-a594-dd443cf9eacc-images\") pod \"machine-config-operator-74547568cd-xk69x\" (UID: \"e18d552a-2c5e-4671-a594-dd443cf9eacc\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xk69x" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.639737 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.658940 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.667019 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e18d552a-2c5e-4671-a594-dd443cf9eacc-proxy-tls\") pod \"machine-config-operator-74547568cd-xk69x\" (UID: \"e18d552a-2c5e-4671-a594-dd443cf9eacc\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xk69x" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.681957 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.690679 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e886b855-272d-435e-ae5b-1498f637409d-config-volume\") pod \"dns-default-4bdpn\" (UID: \"e886b855-272d-435e-ae5b-1498f637409d\") " pod="openshift-dns/dns-default-4bdpn" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.699298 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.719430 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.728651 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e886b855-272d-435e-ae5b-1498f637409d-metrics-tls\") pod \"dns-default-4bdpn\" (UID: \"e886b855-272d-435e-ae5b-1498f637409d\") " pod="openshift-dns/dns-default-4bdpn" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.737507 4733 request.go:700] Waited for 1.922249558s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/hostpath-provisioner/secrets?fieldSelector=metadata.name%3Dcsi-hostpath-provisioner-sa-dockercfg-qd74k&limit=500&resourceVersion=0 Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.740385 4733 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.762905 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.779811 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.818658 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.839956 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.858516 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.878705 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.898651 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.917876 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.938506 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.982588 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlcd5\" (UniqueName: \"kubernetes.io/projected/dec3c184-4636-43b5-9cc4-2188635c2719-kube-api-access-hlcd5\") pod \"etcd-operator-b45778765-t4x4n\" (UID: \"dec3c184-4636-43b5-9cc4-2188635c2719\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t4x4n" Feb 02 15:17:41 crc kubenswrapper[4733]: I0202 15:17:41.987081 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-t4x4n" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.002918 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6cv9\" (UniqueName: \"kubernetes.io/projected/e18d552a-2c5e-4671-a594-dd443cf9eacc-kube-api-access-z6cv9\") pod \"machine-config-operator-74547568cd-xk69x\" (UID: \"e18d552a-2c5e-4671-a594-dd443cf9eacc\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xk69x" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.023319 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lvtx\" (UniqueName: \"kubernetes.io/projected/ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e-kube-api-access-7lvtx\") pod \"marketplace-operator-79b997595-csdnv\" (UID: \"ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e\") " pod="openshift-marketplace/marketplace-operator-79b997595-csdnv" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.035738 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twbv5\" (UniqueName: \"kubernetes.io/projected/613678ea-f8f5-4fc1-beb2-0fde3dfe77ac-kube-api-access-twbv5\") pod \"multus-admission-controller-857f4d67dd-l595n\" (UID: \"613678ea-f8f5-4fc1-beb2-0fde3dfe77ac\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-l595n" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.053285 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjcqw\" (UniqueName: \"kubernetes.io/projected/a0d837f7-b35f-4879-b0f9-d82411bf16ff-kube-api-access-cjcqw\") pod \"packageserver-d55dfcdfc-kxfm5\" (UID: \"a0d837f7-b35f-4879-b0f9-d82411bf16ff\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kxfm5" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.073299 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxzfn\" (UniqueName: \"kubernetes.io/projected/e886b855-272d-435e-ae5b-1498f637409d-kube-api-access-qxzfn\") pod \"dns-default-4bdpn\" (UID: \"e886b855-272d-435e-ae5b-1498f637409d\") " pod="openshift-dns/dns-default-4bdpn" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.075430 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-csdnv" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.079347 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-l595n" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.095767 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kxfm5" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.097777 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkl7b\" (UniqueName: \"kubernetes.io/projected/2c71221f-f379-4474-b1b5-b948c663a1d1-kube-api-access-qkl7b\") pod \"machine-config-controller-84d6567774-l2gg8\" (UID: \"2c71221f-f379-4474-b1b5-b948c663a1d1\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-l2gg8" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.114185 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xk69x" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.118055 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfl9v\" (UniqueName: \"kubernetes.io/projected/c7442bb0-8bba-418d-a9e1-7f935fb07257-kube-api-access-dfl9v\") pod \"package-server-manager-789f6589d5-xhg5n\" (UID: \"c7442bb0-8bba-418d-a9e1-7f935fb07257\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xhg5n" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.119801 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-4bdpn" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.128735 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-r76vv" event={"ID":"56fd5dd1-12e2-4b22-a4ab-291d6bb3ce92","Type":"ContainerStarted","Data":"2805df8f462a25ad07a5069fc1f64ca63de1ed556e7a3f48ba2df467ebe0723e"} Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.128774 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-r76vv" event={"ID":"56fd5dd1-12e2-4b22-a4ab-291d6bb3ce92","Type":"ContainerStarted","Data":"79a0002aaaa9635c16f5ca2ca618ae2d228cda2acfdeb2aeae4b0a12f6bc94f4"} Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.128783 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-r76vv" event={"ID":"56fd5dd1-12e2-4b22-a4ab-291d6bb3ce92","Type":"ContainerStarted","Data":"4a4ea4ca8820d4f471e8c1ec6e26d5d417b4627afa4d727141cd3bc33dcb283d"} Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.134969 4733 generic.go:334] "Generic (PLEG): container finished" podID="a00e530f-3c79-40da-a3f3-faf504af3fea" containerID="f2443320de9e0ec94506dcdff64740c0fa361d8afdee07e07a0d2f7055a48c63" exitCode=0 Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.135017 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gm548" event={"ID":"a00e530f-3c79-40da-a3f3-faf504af3fea","Type":"ContainerDied","Data":"f2443320de9e0ec94506dcdff64740c0fa361d8afdee07e07a0d2f7055a48c63"} Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.143322 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-knzns\" (UniqueName: \"kubernetes.io/projected/f5ff4cdb-523e-49f2-9d0f-54eca688b223-kube-api-access-knzns\") pod \"csi-hostpathplugin-mw8sl\" (UID: \"f5ff4cdb-523e-49f2-9d0f-54eca688b223\") " pod="hostpath-provisioner/csi-hostpathplugin-mw8sl" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.155900 4733 generic.go:334] "Generic (PLEG): container finished" podID="b9dce599-89b8-4c6c-a12c-65f9ef64de9b" containerID="6d44061d4f97a179739796da1639f0eabdb559fa7ecbc7a25fc8b230b563ddaf" exitCode=0 Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.156183 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-wd9qj" event={"ID":"b9dce599-89b8-4c6c-a12c-65f9ef64de9b","Type":"ContainerDied","Data":"6d44061d4f97a179739796da1639f0eabdb559fa7ecbc7a25fc8b230b563ddaf"} Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.172975 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9j9k\" (UniqueName: \"kubernetes.io/projected/a8928569-d787-4825-883d-6adf585f069b-kube-api-access-h9j9k\") pod \"olm-operator-6b444d44fb-wpqs8\" (UID: \"a8928569-d787-4825-883d-6adf585f069b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wpqs8" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.195921 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dgd6\" (UniqueName: \"kubernetes.io/projected/a11da899-1807-4360-be80-ea30f6ef2c5f-kube-api-access-7dgd6\") pod \"catalog-operator-68c6474976-vbpch\" (UID: \"a11da899-1807-4360-be80-ea30f6ef2c5f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vbpch" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.212592 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rdfcg" event={"ID":"4849186f-21ae-4654-a813-4b176093abbd","Type":"ContainerStarted","Data":"1def9245b3f9f5f7a3e3b4adde293652176b7ddcbf8a2449bbadf264c41d43e7"} Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.212632 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rdfcg" event={"ID":"4849186f-21ae-4654-a813-4b176093abbd","Type":"ContainerStarted","Data":"ff4af67ee8dffafbc18846f8c6f681bd308f0d30d82aa4a1ca3c23da2604aea7"} Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.212641 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rdfcg" event={"ID":"4849186f-21ae-4654-a813-4b176093abbd","Type":"ContainerStarted","Data":"9ee504b4ce5cdb2c47168548f0ec970a6d0f256e5ac161c8a3e226715c8fc64d"} Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.222748 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpb9v\" (UniqueName: \"kubernetes.io/projected/b2deaba5-c009-4df6-87b9-185b9a8addc5-kube-api-access-lpb9v\") pod \"service-ca-9c57cc56f-c6qfs\" (UID: \"b2deaba5-c009-4df6-87b9-185b9a8addc5\") " pod="openshift-service-ca/service-ca-9c57cc56f-c6qfs" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.278447 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-4cq4j" event={"ID":"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1","Type":"ContainerStarted","Data":"67433a8e25fd1c4786afeb674ca26aad6693a23348069352b1158a5edc1054a8"} Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.279284 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-4cq4j" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.291727 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-f8l78" event={"ID":"317b7923-187f-47a4-9b82-53b20c86d9a4","Type":"ContainerStarted","Data":"ea998fc38195ea33018608c103f51fe9d5d68abbc77b7f0edfa884cba8096ee6"} Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.303453 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xhg5n" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.305321 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/073bb1a3-d5e7-48ce-9788-b2bc7ee323cc-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-hfp26\" (UID: \"073bb1a3-d5e7-48ce-9788-b2bc7ee323cc\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hfp26" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.305352 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5d46e595-5d24-4e8f-bdf5-32a6bd431def-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-vgtrr\" (UID: \"5d46e595-5d24-4e8f-bdf5-32a6bd431def\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vgtrr" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.305384 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/db551bb0-6793-40de-9b5b-c0ee0dc62711-ca-trust-extracted\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.305415 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/db551bb0-6793-40de-9b5b-c0ee0dc62711-trusted-ca\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.305452 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/db551bb0-6793-40de-9b5b-c0ee0dc62711-bound-sa-token\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.305478 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/02048841-10f9-4ad2-aa2d-c2be2466a950-default-certificate\") pod \"router-default-5444994796-mq75m\" (UID: \"02048841-10f9-4ad2-aa2d-c2be2466a950\") " pod="openshift-ingress/router-default-5444994796-mq75m" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.305541 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/02048841-10f9-4ad2-aa2d-c2be2466a950-stats-auth\") pod \"router-default-5444994796-mq75m\" (UID: \"02048841-10f9-4ad2-aa2d-c2be2466a950\") " pod="openshift-ingress/router-default-5444994796-mq75m" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.305575 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bsxxm\" (UniqueName: \"kubernetes.io/projected/2bbf40f3-3007-4478-a1b5-09774f25b2f7-kube-api-access-bsxxm\") pod \"downloads-7954f5f757-wvgmn\" (UID: \"2bbf40f3-3007-4478-a1b5-09774f25b2f7\") " pod="openshift-console/downloads-7954f5f757-wvgmn" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.305591 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlrpd\" (UniqueName: \"kubernetes.io/projected/93a0b3e7-600b-4a2a-bd8f-04982dd52beb-kube-api-access-tlrpd\") pod \"collect-profiles-29500755-fjlmd\" (UID: \"93a0b3e7-600b-4a2a-bd8f-04982dd52beb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500755-fjlmd" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.305622 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cb61125d-6eee-4362-9aad-02920b017d93-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-pmbhw\" (UID: \"cb61125d-6eee-4362-9aad-02920b017d93\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pmbhw" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.305656 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35f59d89-4814-4c14-aaa2-12837ac4b122-config\") pod \"service-ca-operator-777779d784-gmgdk\" (UID: \"35f59d89-4814-4c14-aaa2-12837ac4b122\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gmgdk" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.305683 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/db551bb0-6793-40de-9b5b-c0ee0dc62711-registry-certificates\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.305757 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5d46e595-5d24-4e8f-bdf5-32a6bd431def-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-vgtrr\" (UID: \"5d46e595-5d24-4e8f-bdf5-32a6bd431def\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vgtrr" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.305813 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/02048841-10f9-4ad2-aa2d-c2be2466a950-service-ca-bundle\") pod \"router-default-5444994796-mq75m\" (UID: \"02048841-10f9-4ad2-aa2d-c2be2466a950\") " pod="openshift-ingress/router-default-5444994796-mq75m" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.305828 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/02048841-10f9-4ad2-aa2d-c2be2466a950-metrics-certs\") pod \"router-default-5444994796-mq75m\" (UID: \"02048841-10f9-4ad2-aa2d-c2be2466a950\") " pod="openshift-ingress/router-default-5444994796-mq75m" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.305845 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mw5h2\" (UniqueName: \"kubernetes.io/projected/db551bb0-6793-40de-9b5b-c0ee0dc62711-kube-api-access-mw5h2\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.305885 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.305941 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/93a0b3e7-600b-4a2a-bd8f-04982dd52beb-config-volume\") pod \"collect-profiles-29500755-fjlmd\" (UID: \"93a0b3e7-600b-4a2a-bd8f-04982dd52beb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500755-fjlmd" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.305966 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/db551bb0-6793-40de-9b5b-c0ee0dc62711-registry-tls\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.305983 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/db551bb0-6793-40de-9b5b-c0ee0dc62711-installation-pull-secrets\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.305998 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hk4zf\" (UniqueName: \"kubernetes.io/projected/02048841-10f9-4ad2-aa2d-c2be2466a950-kube-api-access-hk4zf\") pod \"router-default-5444994796-mq75m\" (UID: \"02048841-10f9-4ad2-aa2d-c2be2466a950\") " pod="openshift-ingress/router-default-5444994796-mq75m" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.306014 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb61125d-6eee-4362-9aad-02920b017d93-config\") pod \"kube-controller-manager-operator-78b949d7b-pmbhw\" (UID: \"cb61125d-6eee-4362-9aad-02920b017d93\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pmbhw" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.306073 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjjxv\" (UniqueName: \"kubernetes.io/projected/4d190cfa-83f0-4044-a011-229aa02d66b9-kube-api-access-gjjxv\") pod \"migrator-59844c95c7-qt46z\" (UID: \"4d190cfa-83f0-4044-a011-229aa02d66b9\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-qt46z" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.306124 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d46e595-5d24-4e8f-bdf5-32a6bd431def-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-vgtrr\" (UID: \"5d46e595-5d24-4e8f-bdf5-32a6bd431def\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vgtrr" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.306192 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89qd6\" (UniqueName: \"kubernetes.io/projected/35f59d89-4814-4c14-aaa2-12837ac4b122-kube-api-access-89qd6\") pod \"service-ca-operator-777779d784-gmgdk\" (UID: \"35f59d89-4814-4c14-aaa2-12837ac4b122\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gmgdk" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.306208 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrtm5\" (UniqueName: \"kubernetes.io/projected/073bb1a3-d5e7-48ce-9788-b2bc7ee323cc-kube-api-access-xrtm5\") pod \"kube-storage-version-migrator-operator-b67b599dd-hfp26\" (UID: \"073bb1a3-d5e7-48ce-9788-b2bc7ee323cc\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hfp26" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.306227 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/93a0b3e7-600b-4a2a-bd8f-04982dd52beb-secret-volume\") pod \"collect-profiles-29500755-fjlmd\" (UID: \"93a0b3e7-600b-4a2a-bd8f-04982dd52beb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500755-fjlmd" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.306242 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/073bb1a3-d5e7-48ce-9788-b2bc7ee323cc-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-hfp26\" (UID: \"073bb1a3-d5e7-48ce-9788-b2bc7ee323cc\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hfp26" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.306256 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/35f59d89-4814-4c14-aaa2-12837ac4b122-serving-cert\") pod \"service-ca-operator-777779d784-gmgdk\" (UID: \"35f59d89-4814-4c14-aaa2-12837ac4b122\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gmgdk" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.306285 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cb61125d-6eee-4362-9aad-02920b017d93-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-pmbhw\" (UID: \"cb61125d-6eee-4362-9aad-02920b017d93\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pmbhw" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.307445 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wpqs8" Feb 02 15:17:42 crc kubenswrapper[4733]: E0202 15:17:42.308325 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 15:17:42.808311279 +0000 UTC m=+146.259772637 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hpgjp" (UID: "db551bb0-6793-40de-9b5b-c0ee0dc62711") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.321061 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hkkwq" event={"ID":"745ab1fe-8d37-420f-9c12-62cd7a33aeaa","Type":"ContainerStarted","Data":"95491ec993181d6f9c93f048c92d476ae74fd3d88fd2aa47b0099498e42784d2"} Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.321101 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hkkwq" event={"ID":"745ab1fe-8d37-420f-9c12-62cd7a33aeaa","Type":"ContainerStarted","Data":"f2b2f2f8cae73d9402cd6291c6a7d8c8692f002ee275b98a7edd1a55b3697bb0"} Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.321286 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-l2gg8" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.324573 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vbpch" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.327623 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-t4x4n"] Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.396363 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-c6qfs" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.397319 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-b68qv" event={"ID":"4e2257ac-9d7a-4e81-b560-32f28a73830e","Type":"ContainerStarted","Data":"f60cbbea39a9098c53a5dd436bce7a9cabc00c871078c0da7c2e039dadca0cbe"} Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.397342 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-b68qv" event={"ID":"4e2257ac-9d7a-4e81-b560-32f28a73830e","Type":"ContainerStarted","Data":"c0db6c1f1192180f915ffb15ee50efdbb205913755e3602724f52c83e013e0dc"} Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.407031 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 15:17:42 crc kubenswrapper[4733]: E0202 15:17:42.407196 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 15:17:42.907151199 +0000 UTC m=+146.358612557 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.407355 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/db551bb0-6793-40de-9b5b-c0ee0dc62711-installation-pull-secrets\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.407425 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hk4zf\" (UniqueName: \"kubernetes.io/projected/02048841-10f9-4ad2-aa2d-c2be2466a950-kube-api-access-hk4zf\") pod \"router-default-5444994796-mq75m\" (UID: \"02048841-10f9-4ad2-aa2d-c2be2466a950\") " pod="openshift-ingress/router-default-5444994796-mq75m" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.407452 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb61125d-6eee-4362-9aad-02920b017d93-config\") pod \"kube-controller-manager-operator-78b949d7b-pmbhw\" (UID: \"cb61125d-6eee-4362-9aad-02920b017d93\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pmbhw" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.407593 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjjxv\" (UniqueName: \"kubernetes.io/projected/4d190cfa-83f0-4044-a011-229aa02d66b9-kube-api-access-gjjxv\") pod \"migrator-59844c95c7-qt46z\" (UID: \"4d190cfa-83f0-4044-a011-229aa02d66b9\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-qt46z" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.407630 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d46e595-5d24-4e8f-bdf5-32a6bd431def-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-vgtrr\" (UID: \"5d46e595-5d24-4e8f-bdf5-32a6bd431def\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vgtrr" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.407714 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mdbl\" (UniqueName: \"kubernetes.io/projected/1d685553-6cbe-4ac6-86c9-bfdd083cd4de-kube-api-access-2mdbl\") pod \"machine-config-server-j6ht9\" (UID: \"1d685553-6cbe-4ac6-86c9-bfdd083cd4de\") " pod="openshift-machine-config-operator/machine-config-server-j6ht9" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.413423 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb61125d-6eee-4362-9aad-02920b017d93-config\") pod \"kube-controller-manager-operator-78b949d7b-pmbhw\" (UID: \"cb61125d-6eee-4362-9aad-02920b017d93\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pmbhw" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.414385 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d46e595-5d24-4e8f-bdf5-32a6bd431def-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-vgtrr\" (UID: \"5d46e595-5d24-4e8f-bdf5-32a6bd431def\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vgtrr" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.427490 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r8cnf" event={"ID":"d667dd66-2215-4d9c-9cf5-64ea2f9a2dc8","Type":"ContainerStarted","Data":"0eacad5d37e65d26855a1ef75075ee87c4ad01c58e3d6f6b2ce1a29ba4f66878"} Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.429063 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89qd6\" (UniqueName: \"kubernetes.io/projected/35f59d89-4814-4c14-aaa2-12837ac4b122-kube-api-access-89qd6\") pod \"service-ca-operator-777779d784-gmgdk\" (UID: \"35f59d89-4814-4c14-aaa2-12837ac4b122\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gmgdk" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.429116 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrtm5\" (UniqueName: \"kubernetes.io/projected/073bb1a3-d5e7-48ce-9788-b2bc7ee323cc-kube-api-access-xrtm5\") pod \"kube-storage-version-migrator-operator-b67b599dd-hfp26\" (UID: \"073bb1a3-d5e7-48ce-9788-b2bc7ee323cc\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hfp26" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.429137 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/93a0b3e7-600b-4a2a-bd8f-04982dd52beb-secret-volume\") pod \"collect-profiles-29500755-fjlmd\" (UID: \"93a0b3e7-600b-4a2a-bd8f-04982dd52beb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500755-fjlmd" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.429156 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/073bb1a3-d5e7-48ce-9788-b2bc7ee323cc-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-hfp26\" (UID: \"073bb1a3-d5e7-48ce-9788-b2bc7ee323cc\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hfp26" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.429181 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/35f59d89-4814-4c14-aaa2-12837ac4b122-serving-cert\") pod \"service-ca-operator-777779d784-gmgdk\" (UID: \"35f59d89-4814-4c14-aaa2-12837ac4b122\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gmgdk" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.429197 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cb61125d-6eee-4362-9aad-02920b017d93-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-pmbhw\" (UID: \"cb61125d-6eee-4362-9aad-02920b017d93\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pmbhw" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.429230 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/073bb1a3-d5e7-48ce-9788-b2bc7ee323cc-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-hfp26\" (UID: \"073bb1a3-d5e7-48ce-9788-b2bc7ee323cc\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hfp26" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.429250 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/1d685553-6cbe-4ac6-86c9-bfdd083cd4de-certs\") pod \"machine-config-server-j6ht9\" (UID: \"1d685553-6cbe-4ac6-86c9-bfdd083cd4de\") " pod="openshift-machine-config-operator/machine-config-server-j6ht9" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.429304 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5d46e595-5d24-4e8f-bdf5-32a6bd431def-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-vgtrr\" (UID: \"5d46e595-5d24-4e8f-bdf5-32a6bd431def\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vgtrr" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.429351 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/db551bb0-6793-40de-9b5b-c0ee0dc62711-ca-trust-extracted\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.429442 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/db551bb0-6793-40de-9b5b-c0ee0dc62711-trusted-ca\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.429485 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/db551bb0-6793-40de-9b5b-c0ee0dc62711-bound-sa-token\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.429545 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/02048841-10f9-4ad2-aa2d-c2be2466a950-default-certificate\") pod \"router-default-5444994796-mq75m\" (UID: \"02048841-10f9-4ad2-aa2d-c2be2466a950\") " pod="openshift-ingress/router-default-5444994796-mq75m" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.429644 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/02048841-10f9-4ad2-aa2d-c2be2466a950-stats-auth\") pod \"router-default-5444994796-mq75m\" (UID: \"02048841-10f9-4ad2-aa2d-c2be2466a950\") " pod="openshift-ingress/router-default-5444994796-mq75m" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.429676 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkkld\" (UniqueName: \"kubernetes.io/projected/5cab73a4-dd7b-45b0-b8a4-972f5359ac63-kube-api-access-rkkld\") pod \"ingress-canary-nzjsl\" (UID: \"5cab73a4-dd7b-45b0-b8a4-972f5359ac63\") " pod="openshift-ingress-canary/ingress-canary-nzjsl" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.429701 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bsxxm\" (UniqueName: \"kubernetes.io/projected/2bbf40f3-3007-4478-a1b5-09774f25b2f7-kube-api-access-bsxxm\") pod \"downloads-7954f5f757-wvgmn\" (UID: \"2bbf40f3-3007-4478-a1b5-09774f25b2f7\") " pod="openshift-console/downloads-7954f5f757-wvgmn" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.429749 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlrpd\" (UniqueName: \"kubernetes.io/projected/93a0b3e7-600b-4a2a-bd8f-04982dd52beb-kube-api-access-tlrpd\") pod \"collect-profiles-29500755-fjlmd\" (UID: \"93a0b3e7-600b-4a2a-bd8f-04982dd52beb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500755-fjlmd" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.429764 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cb61125d-6eee-4362-9aad-02920b017d93-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-pmbhw\" (UID: \"cb61125d-6eee-4362-9aad-02920b017d93\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pmbhw" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.429813 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35f59d89-4814-4c14-aaa2-12837ac4b122-config\") pod \"service-ca-operator-777779d784-gmgdk\" (UID: \"35f59d89-4814-4c14-aaa2-12837ac4b122\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gmgdk" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.429849 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/db551bb0-6793-40de-9b5b-c0ee0dc62711-installation-pull-secrets\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.429875 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/db551bb0-6793-40de-9b5b-c0ee0dc62711-registry-certificates\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.429970 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5d46e595-5d24-4e8f-bdf5-32a6bd431def-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-vgtrr\" (UID: \"5d46e595-5d24-4e8f-bdf5-32a6bd431def\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vgtrr" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.429987 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/02048841-10f9-4ad2-aa2d-c2be2466a950-metrics-certs\") pod \"router-default-5444994796-mq75m\" (UID: \"02048841-10f9-4ad2-aa2d-c2be2466a950\") " pod="openshift-ingress/router-default-5444994796-mq75m" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.430015 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/02048841-10f9-4ad2-aa2d-c2be2466a950-service-ca-bundle\") pod \"router-default-5444994796-mq75m\" (UID: \"02048841-10f9-4ad2-aa2d-c2be2466a950\") " pod="openshift-ingress/router-default-5444994796-mq75m" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.430052 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mw5h2\" (UniqueName: \"kubernetes.io/projected/db551bb0-6793-40de-9b5b-c0ee0dc62711-kube-api-access-mw5h2\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.430095 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.430134 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/1d685553-6cbe-4ac6-86c9-bfdd083cd4de-node-bootstrap-token\") pod \"machine-config-server-j6ht9\" (UID: \"1d685553-6cbe-4ac6-86c9-bfdd083cd4de\") " pod="openshift-machine-config-operator/machine-config-server-j6ht9" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.430293 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5cab73a4-dd7b-45b0-b8a4-972f5359ac63-cert\") pod \"ingress-canary-nzjsl\" (UID: \"5cab73a4-dd7b-45b0-b8a4-972f5359ac63\") " pod="openshift-ingress-canary/ingress-canary-nzjsl" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.430325 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/93a0b3e7-600b-4a2a-bd8f-04982dd52beb-config-volume\") pod \"collect-profiles-29500755-fjlmd\" (UID: \"93a0b3e7-600b-4a2a-bd8f-04982dd52beb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500755-fjlmd" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.430350 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/db551bb0-6793-40de-9b5b-c0ee0dc62711-registry-tls\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.430903 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/073bb1a3-d5e7-48ce-9788-b2bc7ee323cc-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-hfp26\" (UID: \"073bb1a3-d5e7-48ce-9788-b2bc7ee323cc\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hfp26" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.437819 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/02048841-10f9-4ad2-aa2d-c2be2466a950-service-ca-bundle\") pod \"router-default-5444994796-mq75m\" (UID: \"02048841-10f9-4ad2-aa2d-c2be2466a950\") " pod="openshift-ingress/router-default-5444994796-mq75m" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.439557 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/073bb1a3-d5e7-48ce-9788-b2bc7ee323cc-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-hfp26\" (UID: \"073bb1a3-d5e7-48ce-9788-b2bc7ee323cc\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hfp26" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.440415 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/db551bb0-6793-40de-9b5b-c0ee0dc62711-trusted-ca\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:42 crc kubenswrapper[4733]: E0202 15:17:42.441085 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 15:17:42.941069996 +0000 UTC m=+146.392531354 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hpgjp" (UID: "db551bb0-6793-40de-9b5b-c0ee0dc62711") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.441920 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/db551bb0-6793-40de-9b5b-c0ee0dc62711-ca-trust-extracted\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.442353 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-mw8sl" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.443004 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/93a0b3e7-600b-4a2a-bd8f-04982dd52beb-secret-volume\") pod \"collect-profiles-29500755-fjlmd\" (UID: \"93a0b3e7-600b-4a2a-bd8f-04982dd52beb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500755-fjlmd" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.443871 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/db551bb0-6793-40de-9b5b-c0ee0dc62711-registry-certificates\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.444075 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/02048841-10f9-4ad2-aa2d-c2be2466a950-stats-auth\") pod \"router-default-5444994796-mq75m\" (UID: \"02048841-10f9-4ad2-aa2d-c2be2466a950\") " pod="openshift-ingress/router-default-5444994796-mq75m" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.449019 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hk4zf\" (UniqueName: \"kubernetes.io/projected/02048841-10f9-4ad2-aa2d-c2be2466a950-kube-api-access-hk4zf\") pod \"router-default-5444994796-mq75m\" (UID: \"02048841-10f9-4ad2-aa2d-c2be2466a950\") " pod="openshift-ingress/router-default-5444994796-mq75m" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.454332 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/35f59d89-4814-4c14-aaa2-12837ac4b122-serving-cert\") pod \"service-ca-operator-777779d784-gmgdk\" (UID: \"35f59d89-4814-4c14-aaa2-12837ac4b122\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gmgdk" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.456807 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/02048841-10f9-4ad2-aa2d-c2be2466a950-default-certificate\") pod \"router-default-5444994796-mq75m\" (UID: \"02048841-10f9-4ad2-aa2d-c2be2466a950\") " pod="openshift-ingress/router-default-5444994796-mq75m" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.457481 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/02048841-10f9-4ad2-aa2d-c2be2466a950-metrics-certs\") pod \"router-default-5444994796-mq75m\" (UID: \"02048841-10f9-4ad2-aa2d-c2be2466a950\") " pod="openshift-ingress/router-default-5444994796-mq75m" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.461531 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/db551bb0-6793-40de-9b5b-c0ee0dc62711-registry-tls\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.461770 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cb61125d-6eee-4362-9aad-02920b017d93-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-pmbhw\" (UID: \"cb61125d-6eee-4362-9aad-02920b017d93\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pmbhw" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.462706 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/93a0b3e7-600b-4a2a-bd8f-04982dd52beb-config-volume\") pod \"collect-profiles-29500755-fjlmd\" (UID: \"93a0b3e7-600b-4a2a-bd8f-04982dd52beb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500755-fjlmd" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.466582 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35f59d89-4814-4c14-aaa2-12837ac4b122-config\") pod \"service-ca-operator-777779d784-gmgdk\" (UID: \"35f59d89-4814-4c14-aaa2-12837ac4b122\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gmgdk" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.467113 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5d46e595-5d24-4e8f-bdf5-32a6bd431def-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-vgtrr\" (UID: \"5d46e595-5d24-4e8f-bdf5-32a6bd431def\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vgtrr" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.474359 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjjxv\" (UniqueName: \"kubernetes.io/projected/4d190cfa-83f0-4044-a011-229aa02d66b9-kube-api-access-gjjxv\") pod \"migrator-59844c95c7-qt46z\" (UID: \"4d190cfa-83f0-4044-a011-229aa02d66b9\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-qt46z" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.493575 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89qd6\" (UniqueName: \"kubernetes.io/projected/35f59d89-4814-4c14-aaa2-12837ac4b122-kube-api-access-89qd6\") pod \"service-ca-operator-777779d784-gmgdk\" (UID: \"35f59d89-4814-4c14-aaa2-12837ac4b122\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gmgdk" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.526518 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pj8qs" event={"ID":"74e9c10d-c66e-48ea-9813-cb221012dda4","Type":"ContainerStarted","Data":"deb5f5487367558cd58c7265d01a711cbafa47738310f3c0018fa887873a329c"} Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.526563 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pj8qs" event={"ID":"74e9c10d-c66e-48ea-9813-cb221012dda4","Type":"ContainerStarted","Data":"4ca127d0cc07b01584eee5623a64b999e4694ae4841caa367bae3ded2a2ff9c9"} Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.536707 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.543302 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-02 15:12:41 +0000 UTC, rotation deadline is 2026-12-01 11:46:01.720881181 +0000 UTC Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.543331 4733 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 7244h28m19.177552924s for next certificate rotation Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.544477 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/1d685553-6cbe-4ac6-86c9-bfdd083cd4de-certs\") pod \"machine-config-server-j6ht9\" (UID: \"1d685553-6cbe-4ac6-86c9-bfdd083cd4de\") " pod="openshift-machine-config-operator/machine-config-server-j6ht9" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.544610 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkkld\" (UniqueName: \"kubernetes.io/projected/5cab73a4-dd7b-45b0-b8a4-972f5359ac63-kube-api-access-rkkld\") pod \"ingress-canary-nzjsl\" (UID: \"5cab73a4-dd7b-45b0-b8a4-972f5359ac63\") " pod="openshift-ingress-canary/ingress-canary-nzjsl" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.544752 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/1d685553-6cbe-4ac6-86c9-bfdd083cd4de-node-bootstrap-token\") pod \"machine-config-server-j6ht9\" (UID: \"1d685553-6cbe-4ac6-86c9-bfdd083cd4de\") " pod="openshift-machine-config-operator/machine-config-server-j6ht9" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.544791 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5cab73a4-dd7b-45b0-b8a4-972f5359ac63-cert\") pod \"ingress-canary-nzjsl\" (UID: \"5cab73a4-dd7b-45b0-b8a4-972f5359ac63\") " pod="openshift-ingress-canary/ingress-canary-nzjsl" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.544857 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mdbl\" (UniqueName: \"kubernetes.io/projected/1d685553-6cbe-4ac6-86c9-bfdd083cd4de-kube-api-access-2mdbl\") pod \"machine-config-server-j6ht9\" (UID: \"1d685553-6cbe-4ac6-86c9-bfdd083cd4de\") " pod="openshift-machine-config-operator/machine-config-server-j6ht9" Feb 02 15:17:42 crc kubenswrapper[4733]: E0202 15:17:42.545107 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 15:17:43.045091274 +0000 UTC m=+146.496552632 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.558352 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrtm5\" (UniqueName: \"kubernetes.io/projected/073bb1a3-d5e7-48ce-9788-b2bc7ee323cc-kube-api-access-xrtm5\") pod \"kube-storage-version-migrator-operator-b67b599dd-hfp26\" (UID: \"073bb1a3-d5e7-48ce-9788-b2bc7ee323cc\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hfp26" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.570259 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-98h6l" event={"ID":"f7829943-cabc-49c5-a547-62124e38d49e","Type":"ContainerStarted","Data":"918ae4dd606f38b41b783f31ac022be12bb0fe2af0666b1799e8dcec1e790b74"} Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.573706 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-98h6l" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.593243 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-mq75m" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.621045 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-csdnv"] Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.646757 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:42 crc kubenswrapper[4733]: E0202 15:17:42.647618 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 15:17:43.147601143 +0000 UTC m=+146.599062501 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hpgjp" (UID: "db551bb0-6793-40de-9b5b-c0ee0dc62711") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.656785 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-74k4s" event={"ID":"a3e78085-1f5d-4f0c-84bb-62b787b1fbdd","Type":"ContainerStarted","Data":"4462e6087cdb3020e0f07e75b17daedf95fe96be19290083caa64576383f6acd"} Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.656828 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-74k4s" event={"ID":"a3e78085-1f5d-4f0c-84bb-62b787b1fbdd","Type":"ContainerStarted","Data":"1aa8e31d52d15b2427ed90486a0bf08ad02191cc8c678258924affcea59359b0"} Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.661000 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-qt46z" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.667598 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hfp26" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.709404 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-gmgdk" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.712457 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-mt8x9" event={"ID":"a16e3671-5260-4e88-8fcc-8ba2252af6ea","Type":"ContainerStarted","Data":"574c92bfba48ad1e746826e48dc1e11c5987c73ee3b109335fc97a946f4c4600"} Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.716736 4733 generic.go:334] "Generic (PLEG): container finished" podID="2b42593e-4396-4b1a-9272-ca2a2fb5ad6e" containerID="9f283432c1a7078e0716e08358afb0b3536a62f23ac60b2bf158de8bb32e608a" exitCode=0 Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.716808 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l5vld" event={"ID":"2b42593e-4396-4b1a-9272-ca2a2fb5ad6e","Type":"ContainerDied","Data":"9f283432c1a7078e0716e08358afb0b3536a62f23ac60b2bf158de8bb32e608a"} Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.742714 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mw5h2\" (UniqueName: \"kubernetes.io/projected/db551bb0-6793-40de-9b5b-c0ee0dc62711-kube-api-access-mw5h2\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.743070 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/1d685553-6cbe-4ac6-86c9-bfdd083cd4de-node-bootstrap-token\") pod \"machine-config-server-j6ht9\" (UID: \"1d685553-6cbe-4ac6-86c9-bfdd083cd4de\") " pod="openshift-machine-config-operator/machine-config-server-j6ht9" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.743256 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/1d685553-6cbe-4ac6-86c9-bfdd083cd4de-certs\") pod \"machine-config-server-j6ht9\" (UID: \"1d685553-6cbe-4ac6-86c9-bfdd083cd4de\") " pod="openshift-machine-config-operator/machine-config-server-j6ht9" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.743388 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bsxxm\" (UniqueName: \"kubernetes.io/projected/2bbf40f3-3007-4478-a1b5-09774f25b2f7-kube-api-access-bsxxm\") pod \"downloads-7954f5f757-wvgmn\" (UID: \"2bbf40f3-3007-4478-a1b5-09774f25b2f7\") " pod="openshift-console/downloads-7954f5f757-wvgmn" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.743520 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5d46e595-5d24-4e8f-bdf5-32a6bd431def-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-vgtrr\" (UID: \"5d46e595-5d24-4e8f-bdf5-32a6bd431def\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vgtrr" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.747773 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkkld\" (UniqueName: \"kubernetes.io/projected/5cab73a4-dd7b-45b0-b8a4-972f5359ac63-kube-api-access-rkkld\") pod \"ingress-canary-nzjsl\" (UID: \"5cab73a4-dd7b-45b0-b8a4-972f5359ac63\") " pod="openshift-ingress-canary/ingress-canary-nzjsl" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.748225 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.748308 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlrpd\" (UniqueName: \"kubernetes.io/projected/93a0b3e7-600b-4a2a-bd8f-04982dd52beb-kube-api-access-tlrpd\") pod \"collect-profiles-29500755-fjlmd\" (UID: \"93a0b3e7-600b-4a2a-bd8f-04982dd52beb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500755-fjlmd" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.748615 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cb61125d-6eee-4362-9aad-02920b017d93-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-pmbhw\" (UID: \"cb61125d-6eee-4362-9aad-02920b017d93\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pmbhw" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.748729 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/db551bb0-6793-40de-9b5b-c0ee0dc62711-bound-sa-token\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:42 crc kubenswrapper[4733]: E0202 15:17:42.748790 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 15:17:43.248773315 +0000 UTC m=+146.700234673 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.748834 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mdbl\" (UniqueName: \"kubernetes.io/projected/1d685553-6cbe-4ac6-86c9-bfdd083cd4de-kube-api-access-2mdbl\") pod \"machine-config-server-j6ht9\" (UID: \"1d685553-6cbe-4ac6-86c9-bfdd083cd4de\") " pod="openshift-machine-config-operator/machine-config-server-j6ht9" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.749903 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:42 crc kubenswrapper[4733]: E0202 15:17:42.750253 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 15:17:43.250240394 +0000 UTC m=+146.701701752 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hpgjp" (UID: "db551bb0-6793-40de-9b5b-c0ee0dc62711") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.756712 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5cab73a4-dd7b-45b0-b8a4-972f5359ac63-cert\") pod \"ingress-canary-nzjsl\" (UID: \"5cab73a4-dd7b-45b0-b8a4-972f5359ac63\") " pod="openshift-ingress-canary/ingress-canary-nzjsl" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.802515 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dkktg" event={"ID":"bd697754-4524-44ed-be19-18747c0626f5","Type":"ContainerStarted","Data":"9d596bc056d231642bad57583b3f5e6a5f2081853589adccb74285d231e91c0b"} Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.802816 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dkktg" event={"ID":"bd697754-4524-44ed-be19-18747c0626f5","Type":"ContainerStarted","Data":"365892bf6db0aff234f0c5562beb92e8f8f818d51f7165142d77916435f58bf0"} Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.850368 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tdwfd" event={"ID":"3c2b2606-ba50-42a8-ae0b-2176b853cb94","Type":"ContainerStarted","Data":"9e93f267af9eae1a6fceb870392eeeda9246f13b07db0445b890f9e7885095aa"} Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.850402 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tdwfd" event={"ID":"3c2b2606-ba50-42a8-ae0b-2176b853cb94","Type":"ContainerStarted","Data":"a2fd621629c28b37d1e1fa2b4909734d94ccf26ac060b5d02e26c0d99213ce59"} Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.851694 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.860904 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-ct5l6" Feb 02 15:17:42 crc kubenswrapper[4733]: E0202 15:17:42.861182 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 15:17:43.361153097 +0000 UTC m=+146.812614445 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.867292 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-m62b2" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.884466 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pmbhw" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.940007 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kxfm5"] Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.940261 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500755-fjlmd" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.940702 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-wvgmn" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.976687 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-98h6l" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.987359 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.988410 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vgtrr" Feb 02 15:17:42 crc kubenswrapper[4733]: E0202 15:17:42.990645 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 15:17:43.490633586 +0000 UTC m=+146.942094944 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hpgjp" (UID: "db551bb0-6793-40de-9b5b-c0ee0dc62711") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:42 crc kubenswrapper[4733]: I0202 15:17:42.997638 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-4cq4j" Feb 02 15:17:43 crc kubenswrapper[4733]: I0202 15:17:43.042759 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-nzjsl" Feb 02 15:17:43 crc kubenswrapper[4733]: I0202 15:17:43.049488 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-j6ht9" Feb 02 15:17:43 crc kubenswrapper[4733]: I0202 15:17:43.093377 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 15:17:43 crc kubenswrapper[4733]: E0202 15:17:43.093802 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 15:17:43.593788242 +0000 UTC m=+147.045249600 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:43 crc kubenswrapper[4733]: I0202 15:17:43.200979 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:43 crc kubenswrapper[4733]: E0202 15:17:43.201311 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 15:17:43.701299884 +0000 UTC m=+147.152761242 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hpgjp" (UID: "db551bb0-6793-40de-9b5b-c0ee0dc62711") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:43 crc kubenswrapper[4733]: I0202 15:17:43.201836 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-4cq4j" podStartSLOduration=126.201817498 podStartE2EDuration="2m6.201817498s" podCreationTimestamp="2026-02-02 15:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:17:43.201794257 +0000 UTC m=+146.653255615" watchObservedRunningTime="2026-02-02 15:17:43.201817498 +0000 UTC m=+146.653278856" Feb 02 15:17:43 crc kubenswrapper[4733]: I0202 15:17:43.254464 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-l595n"] Feb 02 15:17:43 crc kubenswrapper[4733]: I0202 15:17:43.302369 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-xk69x"] Feb 02 15:17:43 crc kubenswrapper[4733]: I0202 15:17:43.302983 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 15:17:43 crc kubenswrapper[4733]: E0202 15:17:43.303246 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 15:17:43.803232567 +0000 UTC m=+147.254693925 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:43 crc kubenswrapper[4733]: I0202 15:17:43.324904 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-79hfl" podStartSLOduration=127.324888905 podStartE2EDuration="2m7.324888905s" podCreationTimestamp="2026-02-02 15:15:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:17:43.321662919 +0000 UTC m=+146.773124277" watchObservedRunningTime="2026-02-02 15:17:43.324888905 +0000 UTC m=+146.776350263" Feb 02 15:17:43 crc kubenswrapper[4733]: W0202 15:17:43.325416 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1d685553_6cbe_4ac6_86c9_bfdd083cd4de.slice/crio-886a27f0362848180a4a57b8a94c3d9a7e0c44ca85689e6a029155e9872a1ba5 WatchSource:0}: Error finding container 886a27f0362848180a4a57b8a94c3d9a7e0c44ca85689e6a029155e9872a1ba5: Status 404 returned error can't find the container with id 886a27f0362848180a4a57b8a94c3d9a7e0c44ca85689e6a029155e9872a1ba5 Feb 02 15:17:43 crc kubenswrapper[4733]: I0202 15:17:43.392567 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-74k4s" podStartSLOduration=126.392546333 podStartE2EDuration="2m6.392546333s" podCreationTimestamp="2026-02-02 15:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:17:43.362523431 +0000 UTC m=+146.813984789" watchObservedRunningTime="2026-02-02 15:17:43.392546333 +0000 UTC m=+146.844007681" Feb 02 15:17:43 crc kubenswrapper[4733]: I0202 15:17:43.404767 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:43 crc kubenswrapper[4733]: E0202 15:17:43.405090 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 15:17:43.905079937 +0000 UTC m=+147.356541295 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hpgjp" (UID: "db551bb0-6793-40de-9b5b-c0ee0dc62711") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:43 crc kubenswrapper[4733]: I0202 15:17:43.484861 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-m62b2" podStartSLOduration=126.484843527 podStartE2EDuration="2m6.484843527s" podCreationTimestamp="2026-02-02 15:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:17:43.482075703 +0000 UTC m=+146.933537061" watchObservedRunningTime="2026-02-02 15:17:43.484843527 +0000 UTC m=+146.936304885" Feb 02 15:17:43 crc kubenswrapper[4733]: I0202 15:17:43.510608 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 15:17:43 crc kubenswrapper[4733]: E0202 15:17:43.510997 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 15:17:44.010979815 +0000 UTC m=+147.462441173 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:43 crc kubenswrapper[4733]: I0202 15:17:43.615984 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:43 crc kubenswrapper[4733]: E0202 15:17:43.616666 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 15:17:44.116653458 +0000 UTC m=+147.568114816 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hpgjp" (UID: "db551bb0-6793-40de-9b5b-c0ee0dc62711") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:43 crc kubenswrapper[4733]: I0202 15:17:43.631516 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hkkwq" podStartSLOduration=126.631500905 podStartE2EDuration="2m6.631500905s" podCreationTimestamp="2026-02-02 15:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:17:43.588753243 +0000 UTC m=+147.040214601" watchObservedRunningTime="2026-02-02 15:17:43.631500905 +0000 UTC m=+147.082962263" Feb 02 15:17:43 crc kubenswrapper[4733]: I0202 15:17:43.714260 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-r76vv" podStartSLOduration=126.714243465 podStartE2EDuration="2m6.714243465s" podCreationTimestamp="2026-02-02 15:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:17:43.66278189 +0000 UTC m=+147.114243238" watchObservedRunningTime="2026-02-02 15:17:43.714243465 +0000 UTC m=+147.165704823" Feb 02 15:17:43 crc kubenswrapper[4733]: I0202 15:17:43.716951 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 15:17:43 crc kubenswrapper[4733]: E0202 15:17:43.717304 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 15:17:44.217291256 +0000 UTC m=+147.668752614 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:43 crc kubenswrapper[4733]: I0202 15:17:43.781809 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-4bdpn"] Feb 02 15:17:43 crc kubenswrapper[4733]: I0202 15:17:43.799968 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-gmgdk"] Feb 02 15:17:43 crc kubenswrapper[4733]: I0202 15:17:43.818888 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:43 crc kubenswrapper[4733]: E0202 15:17:43.819398 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 15:17:44.319386624 +0000 UTC m=+147.770847982 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hpgjp" (UID: "db551bb0-6793-40de-9b5b-c0ee0dc62711") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:43 crc kubenswrapper[4733]: I0202 15:17:43.845665 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-b68qv" podStartSLOduration=126.845651375 podStartE2EDuration="2m6.845651375s" podCreationTimestamp="2026-02-02 15:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:17:43.843814276 +0000 UTC m=+147.295275634" watchObservedRunningTime="2026-02-02 15:17:43.845651375 +0000 UTC m=+147.297112733" Feb 02 15:17:43 crc kubenswrapper[4733]: W0202 15:17:43.861204 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod35f59d89_4814_4c14_aaa2_12837ac4b122.slice/crio-329c12099e55a9661f3114aff89e41f711867622bf2b6a5bbc837c8ba4a8de6a WatchSource:0}: Error finding container 329c12099e55a9661f3114aff89e41f711867622bf2b6a5bbc837c8ba4a8de6a: Status 404 returned error can't find the container with id 329c12099e55a9661f3114aff89e41f711867622bf2b6a5bbc837c8ba4a8de6a Feb 02 15:17:43 crc kubenswrapper[4733]: I0202 15:17:43.888912 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dkktg" podStartSLOduration=126.888897181 podStartE2EDuration="2m6.888897181s" podCreationTimestamp="2026-02-02 15:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:17:43.887321159 +0000 UTC m=+147.338782517" watchObservedRunningTime="2026-02-02 15:17:43.888897181 +0000 UTC m=+147.340358539" Feb 02 15:17:43 crc kubenswrapper[4733]: I0202 15:17:43.922249 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 15:17:43 crc kubenswrapper[4733]: E0202 15:17:43.922609 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 15:17:44.42257055 +0000 UTC m=+147.874031918 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:43 crc kubenswrapper[4733]: I0202 15:17:43.922732 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:43 crc kubenswrapper[4733]: E0202 15:17:43.923049 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 15:17:44.423042683 +0000 UTC m=+147.874504041 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hpgjp" (UID: "db551bb0-6793-40de-9b5b-c0ee0dc62711") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:43 crc kubenswrapper[4733]: I0202 15:17:43.936203 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-mq75m" event={"ID":"02048841-10f9-4ad2-aa2d-c2be2466a950","Type":"ContainerStarted","Data":"1d91ba4144c79c35370f2c3dd8e0cf907ca2cc9d6a6317adecd7c432399175ae"} Feb 02 15:17:43 crc kubenswrapper[4733]: I0202 15:17:43.936251 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-mq75m" event={"ID":"02048841-10f9-4ad2-aa2d-c2be2466a950","Type":"ContainerStarted","Data":"a56062737a4abfedaea70d9946bcd324752feb6cb33bb04cb23cbb5948b630d6"} Feb 02 15:17:43 crc kubenswrapper[4733]: I0202 15:17:43.950153 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xk69x" event={"ID":"e18d552a-2c5e-4671-a594-dd443cf9eacc","Type":"ContainerStarted","Data":"33890f30ce89916d0a0b95d1d30b911e2e3b8b76c5aa389df0df8bbc6688d214"} Feb 02 15:17:43 crc kubenswrapper[4733]: I0202 15:17:43.958925 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-l2gg8"] Feb 02 15:17:43 crc kubenswrapper[4733]: I0202 15:17:43.971929 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hfp26"] Feb 02 15:17:43 crc kubenswrapper[4733]: I0202 15:17:43.983702 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-j6ht9" event={"ID":"1d685553-6cbe-4ac6-86c9-bfdd083cd4de","Type":"ContainerStarted","Data":"d7c91f2d6b7e78273285c81588ccd4ee5f1bc5061b346f652de674adac2ac99a"} Feb 02 15:17:43 crc kubenswrapper[4733]: I0202 15:17:43.983753 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-j6ht9" event={"ID":"1d685553-6cbe-4ac6-86c9-bfdd083cd4de","Type":"ContainerStarted","Data":"886a27f0362848180a4a57b8a94c3d9a7e0c44ca85689e6a029155e9872a1ba5"} Feb 02 15:17:44 crc kubenswrapper[4733]: I0202 15:17:44.016549 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wpqs8"] Feb 02 15:17:44 crc kubenswrapper[4733]: I0202 15:17:44.023670 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 15:17:44 crc kubenswrapper[4733]: E0202 15:17:44.024787 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 15:17:44.52476077 +0000 UTC m=+147.976222178 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:44 crc kubenswrapper[4733]: I0202 15:17:44.050514 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-mw8sl"] Feb 02 15:17:44 crc kubenswrapper[4733]: I0202 15:17:44.073660 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gm548" event={"ID":"a00e530f-3c79-40da-a3f3-faf504af3fea","Type":"ContainerStarted","Data":"4effe0d15b3a5e15089712155ae3d4163472290071e104c1c10b12c75826cbd6"} Feb 02 15:17:44 crc kubenswrapper[4733]: W0202 15:17:44.074352 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod073bb1a3_d5e7_48ce_9788_b2bc7ee323cc.slice/crio-e67f3ea2a333997313881e20216ee451023fa3c5e3582ec27139977c23a1cefe WatchSource:0}: Error finding container e67f3ea2a333997313881e20216ee451023fa3c5e3582ec27139977c23a1cefe: Status 404 returned error can't find the container with id e67f3ea2a333997313881e20216ee451023fa3c5e3582ec27139977c23a1cefe Feb 02 15:17:44 crc kubenswrapper[4733]: I0202 15:17:44.074603 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gm548" Feb 02 15:17:44 crc kubenswrapper[4733]: I0202 15:17:44.093438 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-wd9qj" event={"ID":"b9dce599-89b8-4c6c-a12c-65f9ef64de9b","Type":"ContainerStarted","Data":"6a9366a6bbe7889dbb817f0a22b548d7ae7257886ff3c35290862d86f6b640e8"} Feb 02 15:17:44 crc kubenswrapper[4733]: I0202 15:17:44.115003 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-l595n" event={"ID":"613678ea-f8f5-4fc1-beb2-0fde3dfe77ac","Type":"ContainerStarted","Data":"c83c71707329ee518b924c42346eb5439a24d0933dd5019a0a9ca716d514e726"} Feb 02 15:17:44 crc kubenswrapper[4733]: I0202 15:17:44.127805 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:44 crc kubenswrapper[4733]: E0202 15:17:44.129129 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 15:17:44.629118118 +0000 UTC m=+148.080579476 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hpgjp" (UID: "db551bb0-6793-40de-9b5b-c0ee0dc62711") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:44 crc kubenswrapper[4733]: I0202 15:17:44.155491 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-t4x4n" event={"ID":"dec3c184-4636-43b5-9cc4-2188635c2719","Type":"ContainerStarted","Data":"828e89c3424ee14134620fdbf6bae0f862be55765c9c8867d68b005acdcf0f37"} Feb 02 15:17:44 crc kubenswrapper[4733]: I0202 15:17:44.155534 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-t4x4n" event={"ID":"dec3c184-4636-43b5-9cc4-2188635c2719","Type":"ContainerStarted","Data":"cfeaeb8f22dd66dc954c2d902f35d7342255192145a5e75d294dd35feb86f1ed"} Feb 02 15:17:44 crc kubenswrapper[4733]: W0202 15:17:44.165265 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda8928569_d787_4825_883d_6adf585f069b.slice/crio-0f2c9a17a07f0ed69350d80470ecf8cda7fee2ec072c597bbb237b0d57560af2 WatchSource:0}: Error finding container 0f2c9a17a07f0ed69350d80470ecf8cda7fee2ec072c597bbb237b0d57560af2: Status 404 returned error can't find the container with id 0f2c9a17a07f0ed69350d80470ecf8cda7fee2ec072c597bbb237b0d57560af2 Feb 02 15:17:44 crc kubenswrapper[4733]: W0202 15:17:44.169435 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf5ff4cdb_523e_49f2_9d0f_54eca688b223.slice/crio-3506d89007312a7998d3c69bc0e1ea7ea534390001168db65103c182cb204b64 WatchSource:0}: Error finding container 3506d89007312a7998d3c69bc0e1ea7ea534390001168db65103c182cb204b64: Status 404 returned error can't find the container with id 3506d89007312a7998d3c69bc0e1ea7ea534390001168db65103c182cb204b64 Feb 02 15:17:44 crc kubenswrapper[4733]: I0202 15:17:44.169499 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kxfm5" event={"ID":"a0d837f7-b35f-4879-b0f9-d82411bf16ff","Type":"ContainerStarted","Data":"3c6391c187ccb25522ef0e390572661ef759e7c1307b3fcc8430cef40cf2e860"} Feb 02 15:17:44 crc kubenswrapper[4733]: I0202 15:17:44.169529 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kxfm5" event={"ID":"a0d837f7-b35f-4879-b0f9-d82411bf16ff","Type":"ContainerStarted","Data":"ce9f7d7d32df83bac6936ba51969dd0a9072fae992d07f8321992e8fec68cb53"} Feb 02 15:17:44 crc kubenswrapper[4733]: I0202 15:17:44.170264 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kxfm5" Feb 02 15:17:44 crc kubenswrapper[4733]: I0202 15:17:44.171598 4733 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-kxfm5 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.39:5443/healthz\": dial tcp 10.217.0.39:5443: connect: connection refused" start-of-body= Feb 02 15:17:44 crc kubenswrapper[4733]: I0202 15:17:44.171639 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kxfm5" podUID="a0d837f7-b35f-4879-b0f9-d82411bf16ff" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.39:5443/healthz\": dial tcp 10.217.0.39:5443: connect: connection refused" Feb 02 15:17:44 crc kubenswrapper[4733]: I0202 15:17:44.176622 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-qt46z"] Feb 02 15:17:44 crc kubenswrapper[4733]: I0202 15:17:44.179110 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-csdnv" event={"ID":"ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e","Type":"ContainerStarted","Data":"35db594c74e3ce27d2532036e1ab3f1815c9fed37332b6cab1836020121772db"} Feb 02 15:17:44 crc kubenswrapper[4733]: I0202 15:17:44.179928 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-csdnv" event={"ID":"ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e","Type":"ContainerStarted","Data":"7e43f999c0810719691f84fda6c80e5d6a519cb83d75ac0f3660b1a7597c4472"} Feb 02 15:17:44 crc kubenswrapper[4733]: I0202 15:17:44.180133 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-csdnv" Feb 02 15:17:44 crc kubenswrapper[4733]: I0202 15:17:44.197883 4733 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-csdnv container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" start-of-body= Feb 02 15:17:44 crc kubenswrapper[4733]: I0202 15:17:44.197922 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-csdnv" podUID="ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" Feb 02 15:17:44 crc kubenswrapper[4733]: I0202 15:17:44.198882 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rdfcg" podStartSLOduration=127.198857921 podStartE2EDuration="2m7.198857921s" podCreationTimestamp="2026-02-02 15:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:17:44.198230634 +0000 UTC m=+147.649691992" watchObservedRunningTime="2026-02-02 15:17:44.198857921 +0000 UTC m=+147.650319279" Feb 02 15:17:44 crc kubenswrapper[4733]: I0202 15:17:44.212100 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l5vld" event={"ID":"2b42593e-4396-4b1a-9272-ca2a2fb5ad6e","Type":"ContainerStarted","Data":"47c160ec7db104942a6e9a24b91e319d961697870a239eb854246182e210c86f"} Feb 02 15:17:44 crc kubenswrapper[4733]: I0202 15:17:44.229295 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 15:17:44 crc kubenswrapper[4733]: E0202 15:17:44.230298 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 15:17:44.73028163 +0000 UTC m=+148.181742988 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:44 crc kubenswrapper[4733]: I0202 15:17:44.247568 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-mt8x9" podStartSLOduration=127.247553232 podStartE2EDuration="2m7.247553232s" podCreationTimestamp="2026-02-02 15:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:17:44.245482366 +0000 UTC m=+147.696943724" watchObservedRunningTime="2026-02-02 15:17:44.247553232 +0000 UTC m=+147.699014590" Feb 02 15:17:44 crc kubenswrapper[4733]: I0202 15:17:44.314022 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-c6qfs"] Feb 02 15:17:44 crc kubenswrapper[4733]: I0202 15:17:44.342645 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:44 crc kubenswrapper[4733]: I0202 15:17:44.342694 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xhg5n"] Feb 02 15:17:44 crc kubenswrapper[4733]: E0202 15:17:44.347351 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 15:17:44.847336227 +0000 UTC m=+148.298797585 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hpgjp" (UID: "db551bb0-6793-40de-9b5b-c0ee0dc62711") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:44 crc kubenswrapper[4733]: I0202 15:17:44.375889 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-f8l78" podStartSLOduration=127.375861159 podStartE2EDuration="2m7.375861159s" podCreationTimestamp="2026-02-02 15:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:17:44.34744156 +0000 UTC m=+147.798902918" watchObservedRunningTime="2026-02-02 15:17:44.375861159 +0000 UTC m=+147.827322517" Feb 02 15:17:44 crc kubenswrapper[4733]: I0202 15:17:44.388548 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-nzjsl"] Feb 02 15:17:44 crc kubenswrapper[4733]: I0202 15:17:44.388587 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vbpch"] Feb 02 15:17:44 crc kubenswrapper[4733]: I0202 15:17:44.402041 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pmbhw"] Feb 02 15:17:44 crc kubenswrapper[4733]: I0202 15:17:44.453976 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 15:17:44 crc kubenswrapper[4733]: E0202 15:17:44.454392 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 15:17:44.954375246 +0000 UTC m=+148.405836594 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:44 crc kubenswrapper[4733]: I0202 15:17:44.520097 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500755-fjlmd"] Feb 02 15:17:44 crc kubenswrapper[4733]: I0202 15:17:44.527928 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-98h6l" podStartSLOduration=127.527912501 podStartE2EDuration="2m7.527912501s" podCreationTimestamp="2026-02-02 15:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:17:44.514430341 +0000 UTC m=+147.965891699" watchObservedRunningTime="2026-02-02 15:17:44.527912501 +0000 UTC m=+147.979373859" Feb 02 15:17:44 crc kubenswrapper[4733]: I0202 15:17:44.562867 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:44 crc kubenswrapper[4733]: E0202 15:17:44.563173 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 15:17:45.063146912 +0000 UTC m=+148.514608270 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hpgjp" (UID: "db551bb0-6793-40de-9b5b-c0ee0dc62711") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:44 crc kubenswrapper[4733]: I0202 15:17:44.581368 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-pj8qs" podStartSLOduration=127.581350768 podStartE2EDuration="2m7.581350768s" podCreationTimestamp="2026-02-02 15:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:17:44.549858017 +0000 UTC m=+148.001319375" watchObservedRunningTime="2026-02-02 15:17:44.581350768 +0000 UTC m=+148.032812116" Feb 02 15:17:44 crc kubenswrapper[4733]: I0202 15:17:44.596784 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-mq75m" Feb 02 15:17:44 crc kubenswrapper[4733]: I0202 15:17:44.601074 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vgtrr"] Feb 02 15:17:44 crc kubenswrapper[4733]: I0202 15:17:44.603087 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-wvgmn"] Feb 02 15:17:44 crc kubenswrapper[4733]: I0202 15:17:44.609339 4733 patch_prober.go:28] interesting pod/router-default-5444994796-mq75m container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 02 15:17:44 crc kubenswrapper[4733]: [-]has-synced failed: reason withheld Feb 02 15:17:44 crc kubenswrapper[4733]: [+]process-running ok Feb 02 15:17:44 crc kubenswrapper[4733]: healthz check failed Feb 02 15:17:44 crc kubenswrapper[4733]: I0202 15:17:44.609382 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mq75m" podUID="02048841-10f9-4ad2-aa2d-c2be2466a950" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 02 15:17:44 crc kubenswrapper[4733]: I0202 15:17:44.663432 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-ct5l6" podStartSLOduration=127.663420081 podStartE2EDuration="2m7.663420081s" podCreationTimestamp="2026-02-02 15:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:17:44.661938631 +0000 UTC m=+148.113399989" watchObservedRunningTime="2026-02-02 15:17:44.663420081 +0000 UTC m=+148.114881429" Feb 02 15:17:44 crc kubenswrapper[4733]: I0202 15:17:44.664265 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 15:17:44 crc kubenswrapper[4733]: E0202 15:17:44.664620 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 15:17:45.164609033 +0000 UTC m=+148.616070391 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:44 crc kubenswrapper[4733]: W0202 15:17:44.713910 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2bbf40f3_3007_4478_a1b5_09774f25b2f7.slice/crio-8d3e47457457e62a5bf4c30da9557a0f5c894237e1828ace82f30359e3a77df6 WatchSource:0}: Error finding container 8d3e47457457e62a5bf4c30da9557a0f5c894237e1828ace82f30359e3a77df6: Status 404 returned error can't find the container with id 8d3e47457457e62a5bf4c30da9557a0f5c894237e1828ace82f30359e3a77df6 Feb 02 15:17:44 crc kubenswrapper[4733]: I0202 15:17:44.767744 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:44 crc kubenswrapper[4733]: E0202 15:17:44.768353 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 15:17:45.268339743 +0000 UTC m=+148.719801101 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hpgjp" (UID: "db551bb0-6793-40de-9b5b-c0ee0dc62711") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:44 crc kubenswrapper[4733]: I0202 15:17:44.802274 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tdwfd" podStartSLOduration=127.80225973 podStartE2EDuration="2m7.80225973s" podCreationTimestamp="2026-02-02 15:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:17:44.769799813 +0000 UTC m=+148.221261171" watchObservedRunningTime="2026-02-02 15:17:44.80225973 +0000 UTC m=+148.253721078" Feb 02 15:17:44 crc kubenswrapper[4733]: I0202 15:17:44.868753 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 15:17:44 crc kubenswrapper[4733]: E0202 15:17:44.869110 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 15:17:45.369091645 +0000 UTC m=+148.820553003 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:44 crc kubenswrapper[4733]: I0202 15:17:44.915998 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r8cnf" podStartSLOduration=127.915979757 podStartE2EDuration="2m7.915979757s" podCreationTimestamp="2026-02-02 15:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:17:44.882478363 +0000 UTC m=+148.333939731" watchObservedRunningTime="2026-02-02 15:17:44.915979757 +0000 UTC m=+148.367441115" Feb 02 15:17:44 crc kubenswrapper[4733]: I0202 15:17:44.970549 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:44 crc kubenswrapper[4733]: E0202 15:17:44.971028 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 15:17:45.471016548 +0000 UTC m=+148.922477906 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hpgjp" (UID: "db551bb0-6793-40de-9b5b-c0ee0dc62711") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:44 crc kubenswrapper[4733]: I0202 15:17:44.983846 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-t4x4n" podStartSLOduration=127.98383251 podStartE2EDuration="2m7.98383251s" podCreationTimestamp="2026-02-02 15:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:17:44.953624773 +0000 UTC m=+148.405086131" watchObservedRunningTime="2026-02-02 15:17:44.98383251 +0000 UTC m=+148.435293868" Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.013481 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kxfm5" podStartSLOduration=128.013462962 podStartE2EDuration="2m8.013462962s" podCreationTimestamp="2026-02-02 15:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:17:44.984137318 +0000 UTC m=+148.435598676" watchObservedRunningTime="2026-02-02 15:17:45.013462962 +0000 UTC m=+148.464924320" Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.032010 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l5vld" podStartSLOduration=128.031991347 podStartE2EDuration="2m8.031991347s" podCreationTimestamp="2026-02-02 15:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:17:45.012301371 +0000 UTC m=+148.463762729" watchObservedRunningTime="2026-02-02 15:17:45.031991347 +0000 UTC m=+148.483452705" Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.074219 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.074770 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:17:45 crc kubenswrapper[4733]: E0202 15:17:45.074981 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 15:17:45.574958104 +0000 UTC m=+149.026419462 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.081374 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-j6ht9" podStartSLOduration=6.081358315 podStartE2EDuration="6.081358315s" podCreationTimestamp="2026-02-02 15:17:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:17:45.034041771 +0000 UTC m=+148.485503129" watchObservedRunningTime="2026-02-02 15:17:45.081358315 +0000 UTC m=+148.532819673" Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.095780 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.114656 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gm548" podStartSLOduration=128.114639944 podStartE2EDuration="2m8.114639944s" podCreationTimestamp="2026-02-02 15:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:17:45.083320388 +0000 UTC m=+148.534781756" watchObservedRunningTime="2026-02-02 15:17:45.114639944 +0000 UTC m=+148.566101292" Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.162347 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-mq75m" podStartSLOduration=128.162330758 podStartE2EDuration="2m8.162330758s" podCreationTimestamp="2026-02-02 15:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:17:45.134202527 +0000 UTC m=+148.585663885" watchObservedRunningTime="2026-02-02 15:17:45.162330758 +0000 UTC m=+148.613792106" Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.178011 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d4dfd44b-b363-4162-9f43-9a8dc4a3b110-metrics-certs\") pod \"network-metrics-daemon-7r9fd\" (UID: \"d4dfd44b-b363-4162-9f43-9a8dc4a3b110\") " pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.178041 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.178062 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.178093 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.178123 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.194560 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.195310 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d4dfd44b-b363-4162-9f43-9a8dc4a3b110-metrics-certs\") pod \"network-metrics-daemon-7r9fd\" (UID: \"d4dfd44b-b363-4162-9f43-9a8dc4a3b110\") " pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.196602 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:17:45 crc kubenswrapper[4733]: E0202 15:17:45.197219 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 15:17:45.69720678 +0000 UTC m=+149.148668138 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hpgjp" (UID: "db551bb0-6793-40de-9b5b-c0ee0dc62711") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.210870 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.299820 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.301156 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 15:17:45 crc kubenswrapper[4733]: E0202 15:17:45.301535 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 15:17:45.801519506 +0000 UTC m=+149.252980864 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.300629 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.305009 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.343588 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xk69x" event={"ID":"e18d552a-2c5e-4671-a594-dd443cf9eacc","Type":"ContainerStarted","Data":"7c56c67cd2e35f7e1637c160f57ef01cc1fcc9a2819157c6e64c404ab7426c5b"} Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.343654 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xk69x" event={"ID":"e18d552a-2c5e-4671-a594-dd443cf9eacc","Type":"ContainerStarted","Data":"57aabf6b4ae7e8b32c84bae0b1117178ffaa1734af5cc65b558cc3863a8bbbb3"} Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.343665 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xhg5n" event={"ID":"c7442bb0-8bba-418d-a9e1-7f935fb07257","Type":"ContainerStarted","Data":"169eccb46e4ae295099d3fdefcd906dd756c450bad7a5c34335b6eaac2262d9b"} Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.343675 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xhg5n" event={"ID":"c7442bb0-8bba-418d-a9e1-7f935fb07257","Type":"ContainerStarted","Data":"fb2fa45a4ce6df97e18edf36fdeb231442f264e72b1bea44d831e1f28f3e399b"} Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.387308 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-4bdpn" event={"ID":"e886b855-272d-435e-ae5b-1498f637409d","Type":"ContainerStarted","Data":"f10db6c2627f7e347bf95652f84402ee089b62327bc168339eb8f10dc574500c"} Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.387369 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-4bdpn" event={"ID":"e886b855-272d-435e-ae5b-1498f637409d","Type":"ContainerStarted","Data":"4ddf32c4d74908d460a46d523f235b98efa957d7712223f9e829b51b3bd27c9c"} Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.387341 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7r9fd" Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.402793 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:45 crc kubenswrapper[4733]: E0202 15:17:45.403098 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 15:17:45.90308718 +0000 UTC m=+149.354548538 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hpgjp" (UID: "db551bb0-6793-40de-9b5b-c0ee0dc62711") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.418530 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500755-fjlmd" event={"ID":"93a0b3e7-600b-4a2a-bd8f-04982dd52beb","Type":"ContainerStarted","Data":"836331d602ddd59bb84c392be4cdaf23827e96a98824ab7b73bf3b09309452e5"} Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.419867 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xk69x" podStartSLOduration=128.419858488 podStartE2EDuration="2m8.419858488s" podCreationTimestamp="2026-02-02 15:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:17:45.419263022 +0000 UTC m=+148.870724380" watchObservedRunningTime="2026-02-02 15:17:45.419858488 +0000 UTC m=+148.871319836" Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.420730 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-csdnv" podStartSLOduration=128.420725941 podStartE2EDuration="2m8.420725941s" podCreationTimestamp="2026-02-02 15:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:17:45.164413584 +0000 UTC m=+148.615874942" watchObservedRunningTime="2026-02-02 15:17:45.420725941 +0000 UTC m=+148.872187299" Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.424422 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l5vld" Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.425012 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l5vld" Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.446580 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-l595n" event={"ID":"613678ea-f8f5-4fc1-beb2-0fde3dfe77ac","Type":"ContainerStarted","Data":"e303e8d137da1e9e5846563bfa2d2610e791337d4c1124198639456199efb04f"} Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.467329 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-qt46z" event={"ID":"4d190cfa-83f0-4044-a011-229aa02d66b9","Type":"ContainerStarted","Data":"5601bbb6ccc8d5df8e420b6313c147ba30a1b1d3088b57bf20fc7ec7deb87227"} Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.467657 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-qt46z" event={"ID":"4d190cfa-83f0-4044-a011-229aa02d66b9","Type":"ContainerStarted","Data":"0a02110bc5d4386106017d01fcdc1aeafc6807ddcdcfd65a4c7fd6acbf32959e"} Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.467667 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-qt46z" event={"ID":"4d190cfa-83f0-4044-a011-229aa02d66b9","Type":"ContainerStarted","Data":"c79ba787efd85fd4b6ced6c02d0ce51346aca5eb1f138c6de1d3ba031988d519"} Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.475621 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29500755-fjlmd" podStartSLOduration=128.475600577 podStartE2EDuration="2m8.475600577s" podCreationTimestamp="2026-02-02 15:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:17:45.448449112 +0000 UTC m=+148.899910470" watchObservedRunningTime="2026-02-02 15:17:45.475600577 +0000 UTC m=+148.927061935" Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.506709 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 15:17:45 crc kubenswrapper[4733]: E0202 15:17:45.506879 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 15:17:46.006856132 +0000 UTC m=+149.458317490 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.507032 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.507451 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-l595n" podStartSLOduration=128.507432057 podStartE2EDuration="2m8.507432057s" podCreationTimestamp="2026-02-02 15:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:17:45.476631834 +0000 UTC m=+148.928093192" watchObservedRunningTime="2026-02-02 15:17:45.507432057 +0000 UTC m=+148.958893415" Feb 02 15:17:45 crc kubenswrapper[4733]: E0202 15:17:45.508047 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 15:17:46.008035843 +0000 UTC m=+149.459497201 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hpgjp" (UID: "db551bb0-6793-40de-9b5b-c0ee0dc62711") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.508457 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-qt46z" podStartSLOduration=128.508453134 podStartE2EDuration="2m8.508453134s" podCreationTimestamp="2026-02-02 15:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:17:45.506679937 +0000 UTC m=+148.958141295" watchObservedRunningTime="2026-02-02 15:17:45.508453134 +0000 UTC m=+148.959914492" Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.512868 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-nzjsl" event={"ID":"5cab73a4-dd7b-45b0-b8a4-972f5359ac63","Type":"ContainerStarted","Data":"4fc8f7d0bfaf22eaf5d7fc66cb2a2711a4afcbf0e5a4fe62675756024dee57d7"} Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.512900 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-nzjsl" event={"ID":"5cab73a4-dd7b-45b0-b8a4-972f5359ac63","Type":"ContainerStarted","Data":"bc5e0243df79cb6559ac0384e2a8b6d40e26d2f4a6ae012527f9a33a54424e7b"} Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.529325 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pmbhw" event={"ID":"cb61125d-6eee-4362-9aad-02920b017d93","Type":"ContainerStarted","Data":"160f80ddba8665b5e5689922135b131102919780fdd1590f4efe2b05c0f1206e"} Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.572475 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l5vld" Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.602819 4733 patch_prober.go:28] interesting pod/router-default-5444994796-mq75m container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 02 15:17:45 crc kubenswrapper[4733]: [-]has-synced failed: reason withheld Feb 02 15:17:45 crc kubenswrapper[4733]: [+]process-running ok Feb 02 15:17:45 crc kubenswrapper[4733]: healthz check failed Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.602885 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mq75m" podUID="02048841-10f9-4ad2-aa2d-c2be2466a950" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.609830 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 15:17:45 crc kubenswrapper[4733]: E0202 15:17:45.610928 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 15:17:46.110913011 +0000 UTC m=+149.562374369 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.614730 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pmbhw" podStartSLOduration=128.614710123 podStartE2EDuration="2m8.614710123s" podCreationTimestamp="2026-02-02 15:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:17:45.612761791 +0000 UTC m=+149.064223149" watchObservedRunningTime="2026-02-02 15:17:45.614710123 +0000 UTC m=+149.066171481" Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.615479 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-nzjsl" podStartSLOduration=6.615472243 podStartE2EDuration="6.615472243s" podCreationTimestamp="2026-02-02 15:17:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:17:45.560015252 +0000 UTC m=+149.011476610" watchObservedRunningTime="2026-02-02 15:17:45.615472243 +0000 UTC m=+149.066933601" Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.665994 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-wd9qj" event={"ID":"b9dce599-89b8-4c6c-a12c-65f9ef64de9b","Type":"ContainerStarted","Data":"d55da9c7f49e50ae8daee1584bf5b74ac80b3262644673d73d4976611a58be51"} Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.696467 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-l2gg8" event={"ID":"2c71221f-f379-4474-b1b5-b948c663a1d1","Type":"ContainerStarted","Data":"d9d438f37359d84c514eec200d17d8efcb183c8a231786fc96e471fd668d6b85"} Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.696780 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-l2gg8" event={"ID":"2c71221f-f379-4474-b1b5-b948c663a1d1","Type":"ContainerStarted","Data":"860f184fa6d25a9aa1d0eeb3ca5ec889405f798b4bb5d661bd13bbf89706e7c4"} Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.696792 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-l2gg8" event={"ID":"2c71221f-f379-4474-b1b5-b948c663a1d1","Type":"ContainerStarted","Data":"4dbec331915e86213f980263411d790d57a532097ef1c73707ab582faf04b0c8"} Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.712383 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:45 crc kubenswrapper[4733]: E0202 15:17:45.712873 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 15:17:46.212862155 +0000 UTC m=+149.664323513 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hpgjp" (UID: "db551bb0-6793-40de-9b5b-c0ee0dc62711") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.721233 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wpqs8" event={"ID":"a8928569-d787-4825-883d-6adf585f069b","Type":"ContainerStarted","Data":"31b825ec0c6353ea8597eeef545cd3d362f7b3f84de0bb9294f003e0e666a086"} Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.721286 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wpqs8" event={"ID":"a8928569-d787-4825-883d-6adf585f069b","Type":"ContainerStarted","Data":"0f2c9a17a07f0ed69350d80470ecf8cda7fee2ec072c597bbb237b0d57560af2"} Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.721795 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wpqs8" Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.726457 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vgtrr" event={"ID":"5d46e595-5d24-4e8f-bdf5-32a6bd431def","Type":"ContainerStarted","Data":"8bddb5374f2397ad2fc28a8bdea9e4fadad4fee4f5c203dd347e01e53fea6fa3"} Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.738733 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hfp26" event={"ID":"073bb1a3-d5e7-48ce-9788-b2bc7ee323cc","Type":"ContainerStarted","Data":"44de701374f654a1e1ab15afca3134792afa19d69927ee1ae5ddc079538de91e"} Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.738774 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hfp26" event={"ID":"073bb1a3-d5e7-48ce-9788-b2bc7ee323cc","Type":"ContainerStarted","Data":"e67f3ea2a333997313881e20216ee451023fa3c5e3582ec27139977c23a1cefe"} Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.740503 4733 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-wpqs8 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" start-of-body= Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.740544 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wpqs8" podUID="a8928569-d787-4825-883d-6adf585f069b" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.741233 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-wd9qj" podStartSLOduration=128.741223622 podStartE2EDuration="2m8.741223622s" podCreationTimestamp="2026-02-02 15:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:17:45.707930133 +0000 UTC m=+149.159391491" watchObservedRunningTime="2026-02-02 15:17:45.741223622 +0000 UTC m=+149.192684980" Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.754012 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-mw8sl" event={"ID":"f5ff4cdb-523e-49f2-9d0f-54eca688b223","Type":"ContainerStarted","Data":"3506d89007312a7998d3c69bc0e1ea7ea534390001168db65103c182cb204b64"} Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.780811 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-l2gg8" podStartSLOduration=128.78079693 podStartE2EDuration="2m8.78079693s" podCreationTimestamp="2026-02-02 15:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:17:45.742592029 +0000 UTC m=+149.194053407" watchObservedRunningTime="2026-02-02 15:17:45.78079693 +0000 UTC m=+149.232258288" Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.782461 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hfp26" podStartSLOduration=128.782455374 podStartE2EDuration="2m8.782455374s" podCreationTimestamp="2026-02-02 15:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:17:45.781374025 +0000 UTC m=+149.232835373" watchObservedRunningTime="2026-02-02 15:17:45.782455374 +0000 UTC m=+149.233916732" Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.812395 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vbpch" event={"ID":"a11da899-1807-4360-be80-ea30f6ef2c5f","Type":"ContainerStarted","Data":"fee8c0460d6f0985c1f9e81a6a5002bdc3a80d74345292cca43a0ad71ef00096"} Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.812572 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vbpch" event={"ID":"a11da899-1807-4360-be80-ea30f6ef2c5f","Type":"ContainerStarted","Data":"bdd15cc1bc44f1cb97760df75873276bcad2947d883a6c6888c299761a028378"} Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.813367 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vbpch" Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.814193 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wpqs8" podStartSLOduration=128.814183981 podStartE2EDuration="2m8.814183981s" podCreationTimestamp="2026-02-02 15:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:17:45.813680878 +0000 UTC m=+149.265142236" watchObservedRunningTime="2026-02-02 15:17:45.814183981 +0000 UTC m=+149.265645339" Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.817137 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 15:17:45 crc kubenswrapper[4733]: E0202 15:17:45.818218 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 15:17:46.318200989 +0000 UTC m=+149.769662347 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.822769 4733 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-vbpch container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.18:8443/healthz\": dial tcp 10.217.0.18:8443: connect: connection refused" start-of-body= Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.822811 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vbpch" podUID="a11da899-1807-4360-be80-ea30f6ef2c5f" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.18:8443/healthz\": dial tcp 10.217.0.18:8443: connect: connection refused" Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.845508 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-wvgmn" event={"ID":"2bbf40f3-3007-4478-a1b5-09774f25b2f7","Type":"ContainerStarted","Data":"8d3e47457457e62a5bf4c30da9557a0f5c894237e1828ace82f30359e3a77df6"} Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.846065 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-wvgmn" Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.848067 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-gmgdk" event={"ID":"35f59d89-4814-4c14-aaa2-12837ac4b122","Type":"ContainerStarted","Data":"2366a301545f7a842cb29722efe9634185fe6aa74efee03a51cfee0fc6780e78"} Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.848190 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-gmgdk" event={"ID":"35f59d89-4814-4c14-aaa2-12837ac4b122","Type":"ContainerStarted","Data":"329c12099e55a9661f3114aff89e41f711867622bf2b6a5bbc837c8ba4a8de6a"} Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.853014 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vbpch" podStartSLOduration=128.852999358 podStartE2EDuration="2m8.852999358s" podCreationTimestamp="2026-02-02 15:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:17:45.85155593 +0000 UTC m=+149.303017288" watchObservedRunningTime="2026-02-02 15:17:45.852999358 +0000 UTC m=+149.304460716" Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.865194 4733 patch_prober.go:28] interesting pod/downloads-7954f5f757-wvgmn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.865379 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-wvgmn" podUID="2bbf40f3-3007-4478-a1b5-09774f25b2f7" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.869265 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-c6qfs" event={"ID":"b2deaba5-c009-4df6-87b9-185b9a8addc5","Type":"ContainerStarted","Data":"f8fb61f7f74cfaacbe4d5934a5caebd64b7e5dd6a116ffdefd66eb5dc2807988"} Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.869332 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-c6qfs" event={"ID":"b2deaba5-c009-4df6-87b9-185b9a8addc5","Type":"ContainerStarted","Data":"28dc404a2f7ab09da2af639fe74f786497a88f902ff979c75d1814bce16e84d6"} Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.871895 4733 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-csdnv container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" start-of-body= Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.871950 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-csdnv" podUID="ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.875745 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-gmgdk" podStartSLOduration=128.875728706 podStartE2EDuration="2m8.875728706s" podCreationTimestamp="2026-02-02 15:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:17:45.873214118 +0000 UTC m=+149.324675476" watchObservedRunningTime="2026-02-02 15:17:45.875728706 +0000 UTC m=+149.327190064" Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.885796 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-l5vld" Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.920820 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:45 crc kubenswrapper[4733]: E0202 15:17:45.921459 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 15:17:46.421443177 +0000 UTC m=+149.872904545 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hpgjp" (UID: "db551bb0-6793-40de-9b5b-c0ee0dc62711") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:45 crc kubenswrapper[4733]: I0202 15:17:45.985384 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-wvgmn" podStartSLOduration=128.985358574 podStartE2EDuration="2m8.985358574s" podCreationTimestamp="2026-02-02 15:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:17:45.902338746 +0000 UTC m=+149.353800104" watchObservedRunningTime="2026-02-02 15:17:45.985358574 +0000 UTC m=+149.436819932" Feb 02 15:17:46 crc kubenswrapper[4733]: I0202 15:17:46.021671 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 15:17:46 crc kubenswrapper[4733]: E0202 15:17:46.022351 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 15:17:46.522327342 +0000 UTC m=+149.973788700 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:46 crc kubenswrapper[4733]: I0202 15:17:46.025086 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:46 crc kubenswrapper[4733]: I0202 15:17:46.030269 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-c6qfs" podStartSLOduration=129.03014287 podStartE2EDuration="2m9.03014287s" podCreationTimestamp="2026-02-02 15:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:17:46.028862496 +0000 UTC m=+149.480323854" watchObservedRunningTime="2026-02-02 15:17:46.03014287 +0000 UTC m=+149.481604228" Feb 02 15:17:46 crc kubenswrapper[4733]: E0202 15:17:46.031666 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 15:17:46.531649331 +0000 UTC m=+149.983110679 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hpgjp" (UID: "db551bb0-6793-40de-9b5b-c0ee0dc62711") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:46 crc kubenswrapper[4733]: I0202 15:17:46.106362 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kxfm5" Feb 02 15:17:46 crc kubenswrapper[4733]: I0202 15:17:46.137487 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 15:17:46 crc kubenswrapper[4733]: E0202 15:17:46.137741 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 15:17:46.637724224 +0000 UTC m=+150.089185582 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:46 crc kubenswrapper[4733]: I0202 15:17:46.239915 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:46 crc kubenswrapper[4733]: E0202 15:17:46.240524 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 15:17:46.74051385 +0000 UTC m=+150.191975208 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hpgjp" (UID: "db551bb0-6793-40de-9b5b-c0ee0dc62711") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:46 crc kubenswrapper[4733]: W0202 15:17:46.258472 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-3798fab1f435e1e268fbf5d044d65e0d107a822bbe58df3813d5b692f38ed1d8 WatchSource:0}: Error finding container 3798fab1f435e1e268fbf5d044d65e0d107a822bbe58df3813d5b692f38ed1d8: Status 404 returned error can't find the container with id 3798fab1f435e1e268fbf5d044d65e0d107a822bbe58df3813d5b692f38ed1d8 Feb 02 15:17:46 crc kubenswrapper[4733]: I0202 15:17:46.341353 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 15:17:46 crc kubenswrapper[4733]: E0202 15:17:46.341598 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 15:17:46.84157764 +0000 UTC m=+150.293038998 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:46 crc kubenswrapper[4733]: W0202 15:17:46.372509 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-772a2736b756a94109f30bdfebb63900b30863a5bfbb99b50c0d1f9f6c00ca10 WatchSource:0}: Error finding container 772a2736b756a94109f30bdfebb63900b30863a5bfbb99b50c0d1f9f6c00ca10: Status 404 returned error can't find the container with id 772a2736b756a94109f30bdfebb63900b30863a5bfbb99b50c0d1f9f6c00ca10 Feb 02 15:17:46 crc kubenswrapper[4733]: I0202 15:17:46.378712 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-7r9fd"] Feb 02 15:17:46 crc kubenswrapper[4733]: W0202 15:17:46.398314 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd4dfd44b_b363_4162_9f43_9a8dc4a3b110.slice/crio-0243ee4a894a334422b06ad5e16048882e4c2b2c5500e527b638bc67330f732d WatchSource:0}: Error finding container 0243ee4a894a334422b06ad5e16048882e4c2b2c5500e527b638bc67330f732d: Status 404 returned error can't find the container with id 0243ee4a894a334422b06ad5e16048882e4c2b2c5500e527b638bc67330f732d Feb 02 15:17:46 crc kubenswrapper[4733]: I0202 15:17:46.442359 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:46 crc kubenswrapper[4733]: E0202 15:17:46.442635 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 15:17:46.942624319 +0000 UTC m=+150.394085677 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hpgjp" (UID: "db551bb0-6793-40de-9b5b-c0ee0dc62711") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:46 crc kubenswrapper[4733]: I0202 15:17:46.534521 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gm548" Feb 02 15:17:46 crc kubenswrapper[4733]: I0202 15:17:46.543791 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 15:17:46 crc kubenswrapper[4733]: E0202 15:17:46.544080 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 15:17:47.044066979 +0000 UTC m=+150.495528337 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:46 crc kubenswrapper[4733]: I0202 15:17:46.608954 4733 patch_prober.go:28] interesting pod/router-default-5444994796-mq75m container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 02 15:17:46 crc kubenswrapper[4733]: [-]has-synced failed: reason withheld Feb 02 15:17:46 crc kubenswrapper[4733]: [+]process-running ok Feb 02 15:17:46 crc kubenswrapper[4733]: healthz check failed Feb 02 15:17:46 crc kubenswrapper[4733]: I0202 15:17:46.609011 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mq75m" podUID="02048841-10f9-4ad2-aa2d-c2be2466a950" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 02 15:17:46 crc kubenswrapper[4733]: I0202 15:17:46.644888 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:46 crc kubenswrapper[4733]: E0202 15:17:46.645237 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 15:17:47.145222551 +0000 UTC m=+150.596683909 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hpgjp" (UID: "db551bb0-6793-40de-9b5b-c0ee0dc62711") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:46 crc kubenswrapper[4733]: I0202 15:17:46.745480 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 15:17:46 crc kubenswrapper[4733]: E0202 15:17:46.745637 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 15:17:47.245614783 +0000 UTC m=+150.697076141 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:46 crc kubenswrapper[4733]: I0202 15:17:46.745811 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:46 crc kubenswrapper[4733]: E0202 15:17:46.746116 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 15:17:47.246107596 +0000 UTC m=+150.697568954 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hpgjp" (UID: "db551bb0-6793-40de-9b5b-c0ee0dc62711") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:46 crc kubenswrapper[4733]: I0202 15:17:46.847188 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 15:17:46 crc kubenswrapper[4733]: E0202 15:17:46.847337 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 15:17:47.34732201 +0000 UTC m=+150.798783368 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:46 crc kubenswrapper[4733]: I0202 15:17:46.847383 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:46 crc kubenswrapper[4733]: E0202 15:17:46.847722 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 15:17:47.3477042 +0000 UTC m=+150.799165558 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hpgjp" (UID: "db551bb0-6793-40de-9b5b-c0ee0dc62711") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:46 crc kubenswrapper[4733]: I0202 15:17:46.884143 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-l595n" event={"ID":"613678ea-f8f5-4fc1-beb2-0fde3dfe77ac","Type":"ContainerStarted","Data":"1ea5665509449dc408c075ef69ec4a1b4c7bb89db958eaf31c921e7b997b27d9"} Feb 02 15:17:46 crc kubenswrapper[4733]: I0202 15:17:46.885981 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-7r9fd" event={"ID":"d4dfd44b-b363-4162-9f43-9a8dc4a3b110","Type":"ContainerStarted","Data":"0243ee4a894a334422b06ad5e16048882e4c2b2c5500e527b638bc67330f732d"} Feb 02 15:17:46 crc kubenswrapper[4733]: I0202 15:17:46.891070 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xhg5n" event={"ID":"c7442bb0-8bba-418d-a9e1-7f935fb07257","Type":"ContainerStarted","Data":"484f55e80a91b16f5454d614d9cdcc405fd066cefe7d5bc46905afb599b553b1"} Feb 02 15:17:46 crc kubenswrapper[4733]: I0202 15:17:46.891388 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xhg5n" Feb 02 15:17:46 crc kubenswrapper[4733]: I0202 15:17:46.898499 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"4aad9de9d7aae613e1125a5ccf29829ec23ec3f4ea01e1d2d48d44ef8c11fb4a"} Feb 02 15:17:46 crc kubenswrapper[4733]: I0202 15:17:46.898534 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"82e23b7ef2d68d15463bb88f1698261bc9358e91a555b6b9996a29b0d217fb33"} Feb 02 15:17:46 crc kubenswrapper[4733]: I0202 15:17:46.899987 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500755-fjlmd" event={"ID":"93a0b3e7-600b-4a2a-bd8f-04982dd52beb","Type":"ContainerStarted","Data":"f841497c36ed2d9978334f785fd8d690e4cafe682319abeee0f66c0988a9f509"} Feb 02 15:17:46 crc kubenswrapper[4733]: I0202 15:17:46.907411 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vgtrr" event={"ID":"5d46e595-5d24-4e8f-bdf5-32a6bd431def","Type":"ContainerStarted","Data":"7c5d4e4299f511754912585c95e0cfabc4089f73a4e916a94253cc3ef7247dc6"} Feb 02 15:17:46 crc kubenswrapper[4733]: I0202 15:17:46.917928 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-mw8sl" event={"ID":"f5ff4cdb-523e-49f2-9d0f-54eca688b223","Type":"ContainerStarted","Data":"28b4bb7ce52bc4172113704dc1e6d9f54d42bfb73252aff7db10de6070985b7b"} Feb 02 15:17:46 crc kubenswrapper[4733]: I0202 15:17:46.924569 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-pmbhw" event={"ID":"cb61125d-6eee-4362-9aad-02920b017d93","Type":"ContainerStarted","Data":"035fc60395a0d751bc42089a142220396da12aaf621ee2fa16d57d594c657327"} Feb 02 15:17:46 crc kubenswrapper[4733]: I0202 15:17:46.945265 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"171d9b4501fa8855cefbb440cb013fff2261c338103e04adcbcdcfede1e589a8"} Feb 02 15:17:46 crc kubenswrapper[4733]: I0202 15:17:46.945333 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"3798fab1f435e1e268fbf5d044d65e0d107a822bbe58df3813d5b692f38ed1d8"} Feb 02 15:17:46 crc kubenswrapper[4733]: I0202 15:17:46.945489 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:17:46 crc kubenswrapper[4733]: I0202 15:17:46.948630 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 15:17:46 crc kubenswrapper[4733]: E0202 15:17:46.949800 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 15:17:47.449785247 +0000 UTC m=+150.901246605 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:46 crc kubenswrapper[4733]: I0202 15:17:46.956978 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-wvgmn" event={"ID":"2bbf40f3-3007-4478-a1b5-09774f25b2f7","Type":"ContainerStarted","Data":"c437880e8bd3d7b21e68b9a6f0c5ddac38a99aae466023374970dcf019315861"} Feb 02 15:17:46 crc kubenswrapper[4733]: I0202 15:17:46.958096 4733 patch_prober.go:28] interesting pod/downloads-7954f5f757-wvgmn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Feb 02 15:17:46 crc kubenswrapper[4733]: I0202 15:17:46.958172 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-wvgmn" podUID="2bbf40f3-3007-4478-a1b5-09774f25b2f7" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" Feb 02 15:17:46 crc kubenswrapper[4733]: I0202 15:17:46.961077 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vgtrr" podStartSLOduration=129.961065798 podStartE2EDuration="2m9.961065798s" podCreationTimestamp="2026-02-02 15:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:17:46.96036149 +0000 UTC m=+150.411822848" watchObservedRunningTime="2026-02-02 15:17:46.961065798 +0000 UTC m=+150.412527156" Feb 02 15:17:46 crc kubenswrapper[4733]: I0202 15:17:46.961112 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"ecea72af22db736be8cf6d9cd551b2848adfb5fd581db8b393bee0288aaebd19"} Feb 02 15:17:46 crc kubenswrapper[4733]: I0202 15:17:46.961141 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"772a2736b756a94109f30bdfebb63900b30863a5bfbb99b50c0d1f9f6c00ca10"} Feb 02 15:17:46 crc kubenswrapper[4733]: I0202 15:17:46.962638 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xhg5n" podStartSLOduration=129.96263359 podStartE2EDuration="2m9.96263359s" podCreationTimestamp="2026-02-02 15:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:17:46.928630402 +0000 UTC m=+150.380091760" watchObservedRunningTime="2026-02-02 15:17:46.96263359 +0000 UTC m=+150.414094948" Feb 02 15:17:46 crc kubenswrapper[4733]: I0202 15:17:46.968813 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-4bdpn" event={"ID":"e886b855-272d-435e-ae5b-1498f637409d","Type":"ContainerStarted","Data":"8f4835b61853048bf08d83c0c6d84bc334178046923af8030382edb6eb71cb77"} Feb 02 15:17:46 crc kubenswrapper[4733]: I0202 15:17:46.970420 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-4bdpn" Feb 02 15:17:46 crc kubenswrapper[4733]: I0202 15:17:46.978847 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-csdnv" Feb 02 15:17:46 crc kubenswrapper[4733]: I0202 15:17:46.989225 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vbpch" Feb 02 15:17:47 crc kubenswrapper[4733]: I0202 15:17:47.001929 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wpqs8" Feb 02 15:17:47 crc kubenswrapper[4733]: I0202 15:17:47.004022 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-4bdpn" podStartSLOduration=8.004011155 podStartE2EDuration="8.004011155s" podCreationTimestamp="2026-02-02 15:17:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:17:47.001653922 +0000 UTC m=+150.453115280" watchObservedRunningTime="2026-02-02 15:17:47.004011155 +0000 UTC m=+150.455472513" Feb 02 15:17:47 crc kubenswrapper[4733]: I0202 15:17:47.050385 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:47 crc kubenswrapper[4733]: E0202 15:17:47.054490 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 15:17:47.554477413 +0000 UTC m=+151.005938771 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hpgjp" (UID: "db551bb0-6793-40de-9b5b-c0ee0dc62711") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:47 crc kubenswrapper[4733]: I0202 15:17:47.151405 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 15:17:47 crc kubenswrapper[4733]: E0202 15:17:47.151783 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 15:17:47.651760572 +0000 UTC m=+151.103221930 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:47 crc kubenswrapper[4733]: I0202 15:17:47.152300 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:47 crc kubenswrapper[4733]: E0202 15:17:47.152640 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 15:17:47.652632295 +0000 UTC m=+151.104093653 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hpgjp" (UID: "db551bb0-6793-40de-9b5b-c0ee0dc62711") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:47 crc kubenswrapper[4733]: I0202 15:17:47.254442 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 15:17:47 crc kubenswrapper[4733]: E0202 15:17:47.254795 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 15:17:47.754765003 +0000 UTC m=+151.206226361 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:47 crc kubenswrapper[4733]: I0202 15:17:47.362034 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:47 crc kubenswrapper[4733]: E0202 15:17:47.362389 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 15:17:47.862366588 +0000 UTC m=+151.313827936 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hpgjp" (UID: "db551bb0-6793-40de-9b5b-c0ee0dc62711") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:47 crc kubenswrapper[4733]: I0202 15:17:47.463214 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 15:17:47 crc kubenswrapper[4733]: E0202 15:17:47.463398 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 15:17:47.963370676 +0000 UTC m=+151.414832034 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:47 crc kubenswrapper[4733]: I0202 15:17:47.463688 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:47 crc kubenswrapper[4733]: E0202 15:17:47.464039 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 15:17:47.964012713 +0000 UTC m=+151.415474071 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hpgjp" (UID: "db551bb0-6793-40de-9b5b-c0ee0dc62711") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:47 crc kubenswrapper[4733]: I0202 15:17:47.564515 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 15:17:47 crc kubenswrapper[4733]: E0202 15:17:47.564722 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 15:17:48.064679272 +0000 UTC m=+151.516140630 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:47 crc kubenswrapper[4733]: I0202 15:17:47.564942 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:47 crc kubenswrapper[4733]: E0202 15:17:47.565266 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 15:17:48.065253087 +0000 UTC m=+151.516714445 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hpgjp" (UID: "db551bb0-6793-40de-9b5b-c0ee0dc62711") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:47 crc kubenswrapper[4733]: I0202 15:17:47.572518 4733 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Feb 02 15:17:47 crc kubenswrapper[4733]: I0202 15:17:47.598757 4733 patch_prober.go:28] interesting pod/router-default-5444994796-mq75m container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 02 15:17:47 crc kubenswrapper[4733]: [-]has-synced failed: reason withheld Feb 02 15:17:47 crc kubenswrapper[4733]: [+]process-running ok Feb 02 15:17:47 crc kubenswrapper[4733]: healthz check failed Feb 02 15:17:47 crc kubenswrapper[4733]: I0202 15:17:47.598822 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mq75m" podUID="02048841-10f9-4ad2-aa2d-c2be2466a950" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 02 15:17:47 crc kubenswrapper[4733]: I0202 15:17:47.666336 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 15:17:47 crc kubenswrapper[4733]: E0202 15:17:47.666519 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 15:17:48.166496552 +0000 UTC m=+151.617957910 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:47 crc kubenswrapper[4733]: I0202 15:17:47.666565 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:47 crc kubenswrapper[4733]: E0202 15:17:47.666984 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 15:17:48.166977095 +0000 UTC m=+151.618438453 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hpgjp" (UID: "db551bb0-6793-40de-9b5b-c0ee0dc62711") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:47 crc kubenswrapper[4733]: I0202 15:17:47.726570 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:17:47 crc kubenswrapper[4733]: I0202 15:17:47.767929 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 15:17:47 crc kubenswrapper[4733]: E0202 15:17:47.768216 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 15:17:48.268203369 +0000 UTC m=+151.719664727 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:47 crc kubenswrapper[4733]: I0202 15:17:47.868999 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:47 crc kubenswrapper[4733]: E0202 15:17:47.870063 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 15:17:48.3700511 +0000 UTC m=+151.821512458 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hpgjp" (UID: "db551bb0-6793-40de-9b5b-c0ee0dc62711") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:47 crc kubenswrapper[4733]: I0202 15:17:47.969673 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 15:17:47 crc kubenswrapper[4733]: E0202 15:17:47.969847 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 15:17:48.469821895 +0000 UTC m=+151.921283253 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:47 crc kubenswrapper[4733]: I0202 15:17:47.974539 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-mw8sl" event={"ID":"f5ff4cdb-523e-49f2-9d0f-54eca688b223","Type":"ContainerStarted","Data":"bfb9c45bc531531dabf42f7b31760091fec2ed9103474aad79330807eca2c613"} Feb 02 15:17:47 crc kubenswrapper[4733]: I0202 15:17:47.974582 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-mw8sl" event={"ID":"f5ff4cdb-523e-49f2-9d0f-54eca688b223","Type":"ContainerStarted","Data":"b364695902cd0b6c2111948d861c30a5b0afae880f4fb7e382650801f58c1c78"} Feb 02 15:17:47 crc kubenswrapper[4733]: I0202 15:17:47.974593 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-mw8sl" event={"ID":"f5ff4cdb-523e-49f2-9d0f-54eca688b223","Type":"ContainerStarted","Data":"5ca6a3746d321ed0e38af4afaeadc577a09e41734b406b1f4901ec12bc0550cf"} Feb 02 15:17:47 crc kubenswrapper[4733]: I0202 15:17:47.976181 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-7r9fd" event={"ID":"d4dfd44b-b363-4162-9f43-9a8dc4a3b110","Type":"ContainerStarted","Data":"5a2e53c1d77865512f7d80bf1b08cea7994bc6a5a87df2ccbe8343ec0c8ed087"} Feb 02 15:17:47 crc kubenswrapper[4733]: I0202 15:17:47.976218 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-7r9fd" event={"ID":"d4dfd44b-b363-4162-9f43-9a8dc4a3b110","Type":"ContainerStarted","Data":"ca057b420ffa795367ab9fd3c14f026e95106c99a6da4407185e444dab3c5381"} Feb 02 15:17:47 crc kubenswrapper[4733]: I0202 15:17:47.977839 4733 patch_prober.go:28] interesting pod/downloads-7954f5f757-wvgmn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Feb 02 15:17:47 crc kubenswrapper[4733]: I0202 15:17:47.977878 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-wvgmn" podUID="2bbf40f3-3007-4478-a1b5-09774f25b2f7" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.027654 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-mw8sl" podStartSLOduration=9.027619599 podStartE2EDuration="9.027619599s" podCreationTimestamp="2026-02-02 15:17:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:17:48.003658819 +0000 UTC m=+151.455120177" watchObservedRunningTime="2026-02-02 15:17:48.027619599 +0000 UTC m=+151.479080957" Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.029384 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-7r9fd" podStartSLOduration=132.029375626 podStartE2EDuration="2m12.029375626s" podCreationTimestamp="2026-02-02 15:15:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:17:48.020147629 +0000 UTC m=+151.471608987" watchObservedRunningTime="2026-02-02 15:17:48.029375626 +0000 UTC m=+151.480836984" Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.071636 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:48 crc kubenswrapper[4733]: E0202 15:17:48.072946 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 15:17:48.572930439 +0000 UTC m=+152.024391887 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hpgjp" (UID: "db551bb0-6793-40de-9b5b-c0ee0dc62711") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.101993 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7gm8t"] Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.102938 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7gm8t" Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.104337 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.120783 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7gm8t"] Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.174896 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 15:17:48 crc kubenswrapper[4733]: E0202 15:17:48.175082 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 15:17:48.675051477 +0000 UTC m=+152.126512835 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.175534 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.175671 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2-catalog-content\") pod \"certified-operators-7gm8t\" (UID: \"d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2\") " pod="openshift-marketplace/certified-operators-7gm8t" Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.175803 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2-utilities\") pod \"certified-operators-7gm8t\" (UID: \"d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2\") " pod="openshift-marketplace/certified-operators-7gm8t" Feb 02 15:17:48 crc kubenswrapper[4733]: E0202 15:17:48.175900 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 15:17:48.675888759 +0000 UTC m=+152.127350197 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hpgjp" (UID: "db551bb0-6793-40de-9b5b-c0ee0dc62711") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.175940 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kvxx\" (UniqueName: \"kubernetes.io/projected/d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2-kube-api-access-2kvxx\") pod \"certified-operators-7gm8t\" (UID: \"d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2\") " pod="openshift-marketplace/certified-operators-7gm8t" Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.277639 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 15:17:48 crc kubenswrapper[4733]: E0202 15:17:48.277792 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 15:17:48.777771011 +0000 UTC m=+152.229232369 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.278275 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2-utilities\") pod \"certified-operators-7gm8t\" (UID: \"d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2\") " pod="openshift-marketplace/certified-operators-7gm8t" Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.278388 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kvxx\" (UniqueName: \"kubernetes.io/projected/d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2-kube-api-access-2kvxx\") pod \"certified-operators-7gm8t\" (UID: \"d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2\") " pod="openshift-marketplace/certified-operators-7gm8t" Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.278505 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.278590 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2-catalog-content\") pod \"certified-operators-7gm8t\" (UID: \"d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2\") " pod="openshift-marketplace/certified-operators-7gm8t" Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.278654 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2-utilities\") pod \"certified-operators-7gm8t\" (UID: \"d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2\") " pod="openshift-marketplace/certified-operators-7gm8t" Feb 02 15:17:48 crc kubenswrapper[4733]: E0202 15:17:48.279012 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-02 15:17:48.779002594 +0000 UTC m=+152.230463952 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hpgjp" (UID: "db551bb0-6793-40de-9b5b-c0ee0dc62711") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.279119 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2-catalog-content\") pod \"certified-operators-7gm8t\" (UID: \"d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2\") " pod="openshift-marketplace/certified-operators-7gm8t" Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.299491 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jk5b2"] Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.300499 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jk5b2" Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.301026 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kvxx\" (UniqueName: \"kubernetes.io/projected/d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2-kube-api-access-2kvxx\") pod \"certified-operators-7gm8t\" (UID: \"d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2\") " pod="openshift-marketplace/certified-operators-7gm8t" Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.303786 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.312500 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jk5b2"] Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.379875 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.380079 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f03c333-e449-4a18-8efb-f8834911a599-utilities\") pod \"community-operators-jk5b2\" (UID: \"4f03c333-e449-4a18-8efb-f8834911a599\") " pod="openshift-marketplace/community-operators-jk5b2" Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.380118 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f03c333-e449-4a18-8efb-f8834911a599-catalog-content\") pod \"community-operators-jk5b2\" (UID: \"4f03c333-e449-4a18-8efb-f8834911a599\") " pod="openshift-marketplace/community-operators-jk5b2" Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.380145 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twtp2\" (UniqueName: \"kubernetes.io/projected/4f03c333-e449-4a18-8efb-f8834911a599-kube-api-access-twtp2\") pod \"community-operators-jk5b2\" (UID: \"4f03c333-e449-4a18-8efb-f8834911a599\") " pod="openshift-marketplace/community-operators-jk5b2" Feb 02 15:17:48 crc kubenswrapper[4733]: E0202 15:17:48.380287 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-02 15:17:48.880273119 +0000 UTC m=+152.331734477 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.420284 4733 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-02-02T15:17:47.572540222Z","Handler":null,"Name":""} Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.422428 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7gm8t" Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.434479 4733 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.434510 4733 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.481530 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f03c333-e449-4a18-8efb-f8834911a599-utilities\") pod \"community-operators-jk5b2\" (UID: \"4f03c333-e449-4a18-8efb-f8834911a599\") " pod="openshift-marketplace/community-operators-jk5b2" Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.481577 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.481600 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f03c333-e449-4a18-8efb-f8834911a599-catalog-content\") pod \"community-operators-jk5b2\" (UID: \"4f03c333-e449-4a18-8efb-f8834911a599\") " pod="openshift-marketplace/community-operators-jk5b2" Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.481638 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twtp2\" (UniqueName: \"kubernetes.io/projected/4f03c333-e449-4a18-8efb-f8834911a599-kube-api-access-twtp2\") pod \"community-operators-jk5b2\" (UID: \"4f03c333-e449-4a18-8efb-f8834911a599\") " pod="openshift-marketplace/community-operators-jk5b2" Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.482342 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f03c333-e449-4a18-8efb-f8834911a599-utilities\") pod \"community-operators-jk5b2\" (UID: \"4f03c333-e449-4a18-8efb-f8834911a599\") " pod="openshift-marketplace/community-operators-jk5b2" Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.482862 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f03c333-e449-4a18-8efb-f8834911a599-catalog-content\") pod \"community-operators-jk5b2\" (UID: \"4f03c333-e449-4a18-8efb-f8834911a599\") " pod="openshift-marketplace/community-operators-jk5b2" Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.485240 4733 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.485272 4733 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.494949 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-m6bvv"] Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.496207 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m6bvv" Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.500651 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-m6bvv"] Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.505264 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twtp2\" (UniqueName: \"kubernetes.io/projected/4f03c333-e449-4a18-8efb-f8834911a599-kube-api-access-twtp2\") pod \"community-operators-jk5b2\" (UID: \"4f03c333-e449-4a18-8efb-f8834911a599\") " pod="openshift-marketplace/community-operators-jk5b2" Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.544756 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hpgjp\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.573180 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.585150 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.585453 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f14a53f-b3e0-46ab-af53-2dfb4041498d-utilities\") pod \"certified-operators-m6bvv\" (UID: \"5f14a53f-b3e0-46ab-af53-2dfb4041498d\") " pod="openshift-marketplace/certified-operators-m6bvv" Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.585518 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d79mk\" (UniqueName: \"kubernetes.io/projected/5f14a53f-b3e0-46ab-af53-2dfb4041498d-kube-api-access-d79mk\") pod \"certified-operators-m6bvv\" (UID: \"5f14a53f-b3e0-46ab-af53-2dfb4041498d\") " pod="openshift-marketplace/certified-operators-m6bvv" Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.585549 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f14a53f-b3e0-46ab-af53-2dfb4041498d-catalog-content\") pod \"certified-operators-m6bvv\" (UID: \"5f14a53f-b3e0-46ab-af53-2dfb4041498d\") " pod="openshift-marketplace/certified-operators-m6bvv" Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.599141 4733 patch_prober.go:28] interesting pod/router-default-5444994796-mq75m container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 02 15:17:48 crc kubenswrapper[4733]: [-]has-synced failed: reason withheld Feb 02 15:17:48 crc kubenswrapper[4733]: [+]process-running ok Feb 02 15:17:48 crc kubenswrapper[4733]: healthz check failed Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.599247 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mq75m" podUID="02048841-10f9-4ad2-aa2d-c2be2466a950" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.610381 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.641025 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jk5b2" Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.682888 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7gm8t"] Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.686942 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f14a53f-b3e0-46ab-af53-2dfb4041498d-catalog-content\") pod \"certified-operators-m6bvv\" (UID: \"5f14a53f-b3e0-46ab-af53-2dfb4041498d\") " pod="openshift-marketplace/certified-operators-m6bvv" Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.687027 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f14a53f-b3e0-46ab-af53-2dfb4041498d-utilities\") pod \"certified-operators-m6bvv\" (UID: \"5f14a53f-b3e0-46ab-af53-2dfb4041498d\") " pod="openshift-marketplace/certified-operators-m6bvv" Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.687067 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d79mk\" (UniqueName: \"kubernetes.io/projected/5f14a53f-b3e0-46ab-af53-2dfb4041498d-kube-api-access-d79mk\") pod \"certified-operators-m6bvv\" (UID: \"5f14a53f-b3e0-46ab-af53-2dfb4041498d\") " pod="openshift-marketplace/certified-operators-m6bvv" Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.688234 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f14a53f-b3e0-46ab-af53-2dfb4041498d-catalog-content\") pod \"certified-operators-m6bvv\" (UID: \"5f14a53f-b3e0-46ab-af53-2dfb4041498d\") " pod="openshift-marketplace/certified-operators-m6bvv" Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.688493 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f14a53f-b3e0-46ab-af53-2dfb4041498d-utilities\") pod \"certified-operators-m6bvv\" (UID: \"5f14a53f-b3e0-46ab-af53-2dfb4041498d\") " pod="openshift-marketplace/certified-operators-m6bvv" Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.694176 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-wmd4l"] Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.695047 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wmd4l" Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.705679 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wmd4l"] Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.724276 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d79mk\" (UniqueName: \"kubernetes.io/projected/5f14a53f-b3e0-46ab-af53-2dfb4041498d-kube-api-access-d79mk\") pod \"certified-operators-m6bvv\" (UID: \"5f14a53f-b3e0-46ab-af53-2dfb4041498d\") " pod="openshift-marketplace/certified-operators-m6bvv" Feb 02 15:17:48 crc kubenswrapper[4733]: W0202 15:17:48.754776 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd9d9f8ab_a82c_4994_a1fd_b62d73eb48d2.slice/crio-e71d638d6e432213eeabe61563af42a90b1c58fca72b381e6d73aa56388e0832 WatchSource:0}: Error finding container e71d638d6e432213eeabe61563af42a90b1c58fca72b381e6d73aa56388e0832: Status 404 returned error can't find the container with id e71d638d6e432213eeabe61563af42a90b1c58fca72b381e6d73aa56388e0832 Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.791832 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jwh9\" (UniqueName: \"kubernetes.io/projected/d116a9fc-670f-4a77-b9ce-d6903d2c3d24-kube-api-access-9jwh9\") pod \"community-operators-wmd4l\" (UID: \"d116a9fc-670f-4a77-b9ce-d6903d2c3d24\") " pod="openshift-marketplace/community-operators-wmd4l" Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.791927 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d116a9fc-670f-4a77-b9ce-d6903d2c3d24-catalog-content\") pod \"community-operators-wmd4l\" (UID: \"d116a9fc-670f-4a77-b9ce-d6903d2c3d24\") " pod="openshift-marketplace/community-operators-wmd4l" Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.791998 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d116a9fc-670f-4a77-b9ce-d6903d2c3d24-utilities\") pod \"community-operators-wmd4l\" (UID: \"d116a9fc-670f-4a77-b9ce-d6903d2c3d24\") " pod="openshift-marketplace/community-operators-wmd4l" Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.852821 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-hpgjp"] Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.853406 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m6bvv" Feb 02 15:17:48 crc kubenswrapper[4733]: W0202 15:17:48.862401 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddb551bb0_6793_40de_9b5b_c0ee0dc62711.slice/crio-29f18ceaa31f8bf05fff8d43cd439a5b7380e1ea586c593065c2aa7a1adf1de2 WatchSource:0}: Error finding container 29f18ceaa31f8bf05fff8d43cd439a5b7380e1ea586c593065c2aa7a1adf1de2: Status 404 returned error can't find the container with id 29f18ceaa31f8bf05fff8d43cd439a5b7380e1ea586c593065c2aa7a1adf1de2 Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.893836 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d116a9fc-670f-4a77-b9ce-d6903d2c3d24-catalog-content\") pod \"community-operators-wmd4l\" (UID: \"d116a9fc-670f-4a77-b9ce-d6903d2c3d24\") " pod="openshift-marketplace/community-operators-wmd4l" Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.893907 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d116a9fc-670f-4a77-b9ce-d6903d2c3d24-utilities\") pod \"community-operators-wmd4l\" (UID: \"d116a9fc-670f-4a77-b9ce-d6903d2c3d24\") " pod="openshift-marketplace/community-operators-wmd4l" Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.893983 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jwh9\" (UniqueName: \"kubernetes.io/projected/d116a9fc-670f-4a77-b9ce-d6903d2c3d24-kube-api-access-9jwh9\") pod \"community-operators-wmd4l\" (UID: \"d116a9fc-670f-4a77-b9ce-d6903d2c3d24\") " pod="openshift-marketplace/community-operators-wmd4l" Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.895954 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d116a9fc-670f-4a77-b9ce-d6903d2c3d24-utilities\") pod \"community-operators-wmd4l\" (UID: \"d116a9fc-670f-4a77-b9ce-d6903d2c3d24\") " pod="openshift-marketplace/community-operators-wmd4l" Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.896202 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d116a9fc-670f-4a77-b9ce-d6903d2c3d24-catalog-content\") pod \"community-operators-wmd4l\" (UID: \"d116a9fc-670f-4a77-b9ce-d6903d2c3d24\") " pod="openshift-marketplace/community-operators-wmd4l" Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.921363 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jwh9\" (UniqueName: \"kubernetes.io/projected/d116a9fc-670f-4a77-b9ce-d6903d2c3d24-kube-api-access-9jwh9\") pod \"community-operators-wmd4l\" (UID: \"d116a9fc-670f-4a77-b9ce-d6903d2c3d24\") " pod="openshift-marketplace/community-operators-wmd4l" Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.965074 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jk5b2"] Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.982679 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" event={"ID":"db551bb0-6793-40de-9b5b-c0ee0dc62711","Type":"ContainerStarted","Data":"29f18ceaa31f8bf05fff8d43cd439a5b7380e1ea586c593065c2aa7a1adf1de2"} Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.984684 4733 generic.go:334] "Generic (PLEG): container finished" podID="d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2" containerID="7c6ed31ff0747875a20b9c5efbf5918d9c17502ec571eca60f281c6b387e8560" exitCode=0 Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.984823 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7gm8t" event={"ID":"d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2","Type":"ContainerDied","Data":"7c6ed31ff0747875a20b9c5efbf5918d9c17502ec571eca60f281c6b387e8560"} Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.984848 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7gm8t" event={"ID":"d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2","Type":"ContainerStarted","Data":"e71d638d6e432213eeabe61563af42a90b1c58fca72b381e6d73aa56388e0832"} Feb 02 15:17:48 crc kubenswrapper[4733]: I0202 15:17:48.986095 4733 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 02 15:17:49 crc kubenswrapper[4733]: I0202 15:17:49.011203 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wmd4l" Feb 02 15:17:49 crc kubenswrapper[4733]: I0202 15:17:49.242226 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wmd4l"] Feb 02 15:17:49 crc kubenswrapper[4733]: I0202 15:17:49.261770 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Feb 02 15:17:49 crc kubenswrapper[4733]: W0202 15:17:49.269679 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd116a9fc_670f_4a77_b9ce_d6903d2c3d24.slice/crio-9e692b73f4ff219635cbbe4910a8408f697662bc9f48fe6ca665875ae99b1e49 WatchSource:0}: Error finding container 9e692b73f4ff219635cbbe4910a8408f697662bc9f48fe6ca665875ae99b1e49: Status 404 returned error can't find the container with id 9e692b73f4ff219635cbbe4910a8408f697662bc9f48fe6ca665875ae99b1e49 Feb 02 15:17:49 crc kubenswrapper[4733]: I0202 15:17:49.370351 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-m6bvv"] Feb 02 15:17:49 crc kubenswrapper[4733]: W0202 15:17:49.383412 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5f14a53f_b3e0_46ab_af53_2dfb4041498d.slice/crio-1d1977da45425a58e7b2201f060d13186326918641a6805bbd30bd6d33494ca0 WatchSource:0}: Error finding container 1d1977da45425a58e7b2201f060d13186326918641a6805bbd30bd6d33494ca0: Status 404 returned error can't find the container with id 1d1977da45425a58e7b2201f060d13186326918641a6805bbd30bd6d33494ca0 Feb 02 15:17:49 crc kubenswrapper[4733]: I0202 15:17:49.597708 4733 patch_prober.go:28] interesting pod/router-default-5444994796-mq75m container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 02 15:17:49 crc kubenswrapper[4733]: [-]has-synced failed: reason withheld Feb 02 15:17:49 crc kubenswrapper[4733]: [+]process-running ok Feb 02 15:17:49 crc kubenswrapper[4733]: healthz check failed Feb 02 15:17:49 crc kubenswrapper[4733]: I0202 15:17:49.597766 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mq75m" podUID="02048841-10f9-4ad2-aa2d-c2be2466a950" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 02 15:17:49 crc kubenswrapper[4733]: I0202 15:17:49.912899 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 02 15:17:49 crc kubenswrapper[4733]: I0202 15:17:49.913686 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 02 15:17:49 crc kubenswrapper[4733]: I0202 15:17:49.915655 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 02 15:17:49 crc kubenswrapper[4733]: I0202 15:17:49.917147 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 02 15:17:49 crc kubenswrapper[4733]: I0202 15:17:49.920411 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.007762 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m6bvv" event={"ID":"5f14a53f-b3e0-46ab-af53-2dfb4041498d","Type":"ContainerDied","Data":"81289aa37886a3f159d89678d35ecbb761830f4804aeb83caf3430344d5ef808"} Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.008236 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fe1172c1-0bf0-475a-b590-24c573598a08-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"fe1172c1-0bf0-475a-b590-24c573598a08\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.008939 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fe1172c1-0bf0-475a-b590-24c573598a08-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"fe1172c1-0bf0-475a-b590-24c573598a08\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.010309 4733 generic.go:334] "Generic (PLEG): container finished" podID="5f14a53f-b3e0-46ab-af53-2dfb4041498d" containerID="81289aa37886a3f159d89678d35ecbb761830f4804aeb83caf3430344d5ef808" exitCode=0 Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.010441 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m6bvv" event={"ID":"5f14a53f-b3e0-46ab-af53-2dfb4041498d","Type":"ContainerStarted","Data":"1d1977da45425a58e7b2201f060d13186326918641a6805bbd30bd6d33494ca0"} Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.014662 4733 generic.go:334] "Generic (PLEG): container finished" podID="4f03c333-e449-4a18-8efb-f8834911a599" containerID="85556a9a2478a4e15cd0900c51c32868a05b1d6954d5ce370f83dba7566ff3e4" exitCode=0 Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.015089 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jk5b2" event={"ID":"4f03c333-e449-4a18-8efb-f8834911a599","Type":"ContainerDied","Data":"85556a9a2478a4e15cd0900c51c32868a05b1d6954d5ce370f83dba7566ff3e4"} Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.015190 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jk5b2" event={"ID":"4f03c333-e449-4a18-8efb-f8834911a599","Type":"ContainerStarted","Data":"751df3ff976e06527b60c7efd19e9458f338ee5a7817ab06b03af72f184f341e"} Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.019664 4733 generic.go:334] "Generic (PLEG): container finished" podID="d116a9fc-670f-4a77-b9ce-d6903d2c3d24" containerID="c1d608ae3eb57339ae0b401ec54724b0ec2601dc46b6e145a1a425f431eab741" exitCode=0 Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.020059 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wmd4l" event={"ID":"d116a9fc-670f-4a77-b9ce-d6903d2c3d24","Type":"ContainerDied","Data":"c1d608ae3eb57339ae0b401ec54724b0ec2601dc46b6e145a1a425f431eab741"} Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.020118 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wmd4l" event={"ID":"d116a9fc-670f-4a77-b9ce-d6903d2c3d24","Type":"ContainerStarted","Data":"9e692b73f4ff219635cbbe4910a8408f697662bc9f48fe6ca665875ae99b1e49"} Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.025209 4733 generic.go:334] "Generic (PLEG): container finished" podID="93a0b3e7-600b-4a2a-bd8f-04982dd52beb" containerID="f841497c36ed2d9978334f785fd8d690e4cafe682319abeee0f66c0988a9f509" exitCode=0 Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.025264 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500755-fjlmd" event={"ID":"93a0b3e7-600b-4a2a-bd8f-04982dd52beb","Type":"ContainerDied","Data":"f841497c36ed2d9978334f785fd8d690e4cafe682319abeee0f66c0988a9f509"} Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.029313 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" event={"ID":"db551bb0-6793-40de-9b5b-c0ee0dc62711","Type":"ContainerStarted","Data":"7980416916a00f1a8903b9ad92053dd587dfeb5605af745c92f56f0db007f90f"} Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.029595 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.082027 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" podStartSLOduration=133.082007568 podStartE2EDuration="2m13.082007568s" podCreationTimestamp="2026-02-02 15:15:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:17:50.076387298 +0000 UTC m=+153.527848676" watchObservedRunningTime="2026-02-02 15:17:50.082007568 +0000 UTC m=+153.533468926" Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.084614 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-wd9qj" Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.084669 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-wd9qj" Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.092684 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-v5t9s"] Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.093715 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v5t9s" Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.099182 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.108121 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v5t9s"] Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.108585 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-wd9qj" Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.110430 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fe1172c1-0bf0-475a-b590-24c573598a08-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"fe1172c1-0bf0-475a-b590-24c573598a08\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.110564 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fe1172c1-0bf0-475a-b590-24c573598a08-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"fe1172c1-0bf0-475a-b590-24c573598a08\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.111664 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fe1172c1-0bf0-475a-b590-24c573598a08-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"fe1172c1-0bf0-475a-b590-24c573598a08\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.136701 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fe1172c1-0bf0-475a-b590-24c573598a08-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"fe1172c1-0bf0-475a-b590-24c573598a08\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.219567 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b6ed8e1-1992-4d83-954a-552aa8e7acb3-utilities\") pod \"redhat-marketplace-v5t9s\" (UID: \"7b6ed8e1-1992-4d83-954a-552aa8e7acb3\") " pod="openshift-marketplace/redhat-marketplace-v5t9s" Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.219628 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b6ed8e1-1992-4d83-954a-552aa8e7acb3-catalog-content\") pod \"redhat-marketplace-v5t9s\" (UID: \"7b6ed8e1-1992-4d83-954a-552aa8e7acb3\") " pod="openshift-marketplace/redhat-marketplace-v5t9s" Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.219747 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7vzn\" (UniqueName: \"kubernetes.io/projected/7b6ed8e1-1992-4d83-954a-552aa8e7acb3-kube-api-access-f7vzn\") pod \"redhat-marketplace-v5t9s\" (UID: \"7b6ed8e1-1992-4d83-954a-552aa8e7acb3\") " pod="openshift-marketplace/redhat-marketplace-v5t9s" Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.266087 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.321206 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7vzn\" (UniqueName: \"kubernetes.io/projected/7b6ed8e1-1992-4d83-954a-552aa8e7acb3-kube-api-access-f7vzn\") pod \"redhat-marketplace-v5t9s\" (UID: \"7b6ed8e1-1992-4d83-954a-552aa8e7acb3\") " pod="openshift-marketplace/redhat-marketplace-v5t9s" Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.321278 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b6ed8e1-1992-4d83-954a-552aa8e7acb3-utilities\") pod \"redhat-marketplace-v5t9s\" (UID: \"7b6ed8e1-1992-4d83-954a-552aa8e7acb3\") " pod="openshift-marketplace/redhat-marketplace-v5t9s" Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.321332 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b6ed8e1-1992-4d83-954a-552aa8e7acb3-catalog-content\") pod \"redhat-marketplace-v5t9s\" (UID: \"7b6ed8e1-1992-4d83-954a-552aa8e7acb3\") " pod="openshift-marketplace/redhat-marketplace-v5t9s" Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.321936 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b6ed8e1-1992-4d83-954a-552aa8e7acb3-catalog-content\") pod \"redhat-marketplace-v5t9s\" (UID: \"7b6ed8e1-1992-4d83-954a-552aa8e7acb3\") " pod="openshift-marketplace/redhat-marketplace-v5t9s" Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.321947 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b6ed8e1-1992-4d83-954a-552aa8e7acb3-utilities\") pod \"redhat-marketplace-v5t9s\" (UID: \"7b6ed8e1-1992-4d83-954a-552aa8e7acb3\") " pod="openshift-marketplace/redhat-marketplace-v5t9s" Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.351763 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7vzn\" (UniqueName: \"kubernetes.io/projected/7b6ed8e1-1992-4d83-954a-552aa8e7acb3-kube-api-access-f7vzn\") pod \"redhat-marketplace-v5t9s\" (UID: \"7b6ed8e1-1992-4d83-954a-552aa8e7acb3\") " pod="openshift-marketplace/redhat-marketplace-v5t9s" Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.413043 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-f8l78" Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.413083 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-f8l78" Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.415884 4733 patch_prober.go:28] interesting pod/console-f9d7485db-f8l78 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.415927 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-f8l78" podUID="317b7923-187f-47a4-9b82-53b20c86d9a4" containerName="console" probeResult="failure" output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.418506 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v5t9s" Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.477972 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.495173 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-76wkk"] Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.502979 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-76wkk" Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.512007 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-76wkk"] Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.596976 4733 patch_prober.go:28] interesting pod/router-default-5444994796-mq75m container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 02 15:17:50 crc kubenswrapper[4733]: [-]has-synced failed: reason withheld Feb 02 15:17:50 crc kubenswrapper[4733]: [+]process-running ok Feb 02 15:17:50 crc kubenswrapper[4733]: healthz check failed Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.597025 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mq75m" podUID="02048841-10f9-4ad2-aa2d-c2be2466a950" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.627835 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/316793cf-5542-4af4-b6ca-ad207ecc5009-catalog-content\") pod \"redhat-marketplace-76wkk\" (UID: \"316793cf-5542-4af4-b6ca-ad207ecc5009\") " pod="openshift-marketplace/redhat-marketplace-76wkk" Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.628037 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/316793cf-5542-4af4-b6ca-ad207ecc5009-utilities\") pod \"redhat-marketplace-76wkk\" (UID: \"316793cf-5542-4af4-b6ca-ad207ecc5009\") " pod="openshift-marketplace/redhat-marketplace-76wkk" Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.628059 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jdp6\" (UniqueName: \"kubernetes.io/projected/316793cf-5542-4af4-b6ca-ad207ecc5009-kube-api-access-4jdp6\") pod \"redhat-marketplace-76wkk\" (UID: \"316793cf-5542-4af4-b6ca-ad207ecc5009\") " pod="openshift-marketplace/redhat-marketplace-76wkk" Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.628615 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v5t9s"] Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.728762 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/316793cf-5542-4af4-b6ca-ad207ecc5009-utilities\") pod \"redhat-marketplace-76wkk\" (UID: \"316793cf-5542-4af4-b6ca-ad207ecc5009\") " pod="openshift-marketplace/redhat-marketplace-76wkk" Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.728797 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jdp6\" (UniqueName: \"kubernetes.io/projected/316793cf-5542-4af4-b6ca-ad207ecc5009-kube-api-access-4jdp6\") pod \"redhat-marketplace-76wkk\" (UID: \"316793cf-5542-4af4-b6ca-ad207ecc5009\") " pod="openshift-marketplace/redhat-marketplace-76wkk" Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.728833 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/316793cf-5542-4af4-b6ca-ad207ecc5009-catalog-content\") pod \"redhat-marketplace-76wkk\" (UID: \"316793cf-5542-4af4-b6ca-ad207ecc5009\") " pod="openshift-marketplace/redhat-marketplace-76wkk" Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.729264 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/316793cf-5542-4af4-b6ca-ad207ecc5009-catalog-content\") pod \"redhat-marketplace-76wkk\" (UID: \"316793cf-5542-4af4-b6ca-ad207ecc5009\") " pod="openshift-marketplace/redhat-marketplace-76wkk" Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.729347 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/316793cf-5542-4af4-b6ca-ad207ecc5009-utilities\") pod \"redhat-marketplace-76wkk\" (UID: \"316793cf-5542-4af4-b6ca-ad207ecc5009\") " pod="openshift-marketplace/redhat-marketplace-76wkk" Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.748389 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jdp6\" (UniqueName: \"kubernetes.io/projected/316793cf-5542-4af4-b6ca-ad207ecc5009-kube-api-access-4jdp6\") pod \"redhat-marketplace-76wkk\" (UID: \"316793cf-5542-4af4-b6ca-ad207ecc5009\") " pod="openshift-marketplace/redhat-marketplace-76wkk" Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.829881 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-76wkk" Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.883951 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.884826 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.890206 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.890564 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Feb 02 15:17:50 crc kubenswrapper[4733]: I0202 15:17:50.894534 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.020961 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-76wkk"] Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.031907 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e92fa8c6-7e1d-4185-896a-e582d573c97d-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e92fa8c6-7e1d-4185-896a-e582d573c97d\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.031985 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e92fa8c6-7e1d-4185-896a-e582d573c97d-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e92fa8c6-7e1d-4185-896a-e582d573c97d\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.043828 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"fe1172c1-0bf0-475a-b590-24c573598a08","Type":"ContainerStarted","Data":"031c19828973b7e3003deccc0f6326036019066a35683f24f1f5c3dca65664dc"} Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.043868 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"fe1172c1-0bf0-475a-b590-24c573598a08","Type":"ContainerStarted","Data":"b382338ee08cdcaeb010e6d398520e352586f8d8b5976bdc02a49a18ae5748a6"} Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.046548 4733 generic.go:334] "Generic (PLEG): container finished" podID="7b6ed8e1-1992-4d83-954a-552aa8e7acb3" containerID="76efd077689c639193102dc831e6ccbaa0f0fdcbeff0a86ff4008a3b266fb8b5" exitCode=0 Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.046715 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v5t9s" event={"ID":"7b6ed8e1-1992-4d83-954a-552aa8e7acb3","Type":"ContainerDied","Data":"76efd077689c639193102dc831e6ccbaa0f0fdcbeff0a86ff4008a3b266fb8b5"} Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.046754 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v5t9s" event={"ID":"7b6ed8e1-1992-4d83-954a-552aa8e7acb3","Type":"ContainerStarted","Data":"6f602392a11e5bfe8ca1c7e652ed4c15847974f2580cb5dfd7dd62af9c7f5dd2"} Feb 02 15:17:51 crc kubenswrapper[4733]: W0202 15:17:51.047484 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod316793cf_5542_4af4_b6ca_ad207ecc5009.slice/crio-a821ee12ee6448197d2deaf68c45ff5902304e01e0569e4a56d06a469dd005d6 WatchSource:0}: Error finding container a821ee12ee6448197d2deaf68c45ff5902304e01e0569e4a56d06a469dd005d6: Status 404 returned error can't find the container with id a821ee12ee6448197d2deaf68c45ff5902304e01e0569e4a56d06a469dd005d6 Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.058950 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.058931134 podStartE2EDuration="2.058931134s" podCreationTimestamp="2026-02-02 15:17:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:17:51.057789704 +0000 UTC m=+154.509251062" watchObservedRunningTime="2026-02-02 15:17:51.058931134 +0000 UTC m=+154.510392492" Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.059091 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-wd9qj" Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.132857 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e92fa8c6-7e1d-4185-896a-e582d573c97d-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e92fa8c6-7e1d-4185-896a-e582d573c97d\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.133047 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e92fa8c6-7e1d-4185-896a-e582d573c97d-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e92fa8c6-7e1d-4185-896a-e582d573c97d\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.134971 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e92fa8c6-7e1d-4185-896a-e582d573c97d-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e92fa8c6-7e1d-4185-896a-e582d573c97d\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.164941 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e92fa8c6-7e1d-4185-896a-e582d573c97d-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e92fa8c6-7e1d-4185-896a-e582d573c97d\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.206694 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.305912 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9fckn"] Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.311327 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9fckn" Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.319144 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.342190 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9fckn"] Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.440272 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52a641c3-5baf-4d1d-a51f-93f372b1013e-catalog-content\") pod \"redhat-operators-9fckn\" (UID: \"52a641c3-5baf-4d1d-a51f-93f372b1013e\") " pod="openshift-marketplace/redhat-operators-9fckn" Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.440510 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52a641c3-5baf-4d1d-a51f-93f372b1013e-utilities\") pod \"redhat-operators-9fckn\" (UID: \"52a641c3-5baf-4d1d-a51f-93f372b1013e\") " pod="openshift-marketplace/redhat-operators-9fckn" Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.440588 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2fgr\" (UniqueName: \"kubernetes.io/projected/52a641c3-5baf-4d1d-a51f-93f372b1013e-kube-api-access-q2fgr\") pod \"redhat-operators-9fckn\" (UID: \"52a641c3-5baf-4d1d-a51f-93f372b1013e\") " pod="openshift-marketplace/redhat-operators-9fckn" Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.466675 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500755-fjlmd" Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.541332 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tlrpd\" (UniqueName: \"kubernetes.io/projected/93a0b3e7-600b-4a2a-bd8f-04982dd52beb-kube-api-access-tlrpd\") pod \"93a0b3e7-600b-4a2a-bd8f-04982dd52beb\" (UID: \"93a0b3e7-600b-4a2a-bd8f-04982dd52beb\") " Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.541374 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/93a0b3e7-600b-4a2a-bd8f-04982dd52beb-config-volume\") pod \"93a0b3e7-600b-4a2a-bd8f-04982dd52beb\" (UID: \"93a0b3e7-600b-4a2a-bd8f-04982dd52beb\") " Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.541409 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/93a0b3e7-600b-4a2a-bd8f-04982dd52beb-secret-volume\") pod \"93a0b3e7-600b-4a2a-bd8f-04982dd52beb\" (UID: \"93a0b3e7-600b-4a2a-bd8f-04982dd52beb\") " Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.541636 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2fgr\" (UniqueName: \"kubernetes.io/projected/52a641c3-5baf-4d1d-a51f-93f372b1013e-kube-api-access-q2fgr\") pod \"redhat-operators-9fckn\" (UID: \"52a641c3-5baf-4d1d-a51f-93f372b1013e\") " pod="openshift-marketplace/redhat-operators-9fckn" Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.541667 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52a641c3-5baf-4d1d-a51f-93f372b1013e-catalog-content\") pod \"redhat-operators-9fckn\" (UID: \"52a641c3-5baf-4d1d-a51f-93f372b1013e\") " pod="openshift-marketplace/redhat-operators-9fckn" Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.541708 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52a641c3-5baf-4d1d-a51f-93f372b1013e-utilities\") pod \"redhat-operators-9fckn\" (UID: \"52a641c3-5baf-4d1d-a51f-93f372b1013e\") " pod="openshift-marketplace/redhat-operators-9fckn" Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.542120 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52a641c3-5baf-4d1d-a51f-93f372b1013e-utilities\") pod \"redhat-operators-9fckn\" (UID: \"52a641c3-5baf-4d1d-a51f-93f372b1013e\") " pod="openshift-marketplace/redhat-operators-9fckn" Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.542279 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93a0b3e7-600b-4a2a-bd8f-04982dd52beb-config-volume" (OuterVolumeSpecName: "config-volume") pod "93a0b3e7-600b-4a2a-bd8f-04982dd52beb" (UID: "93a0b3e7-600b-4a2a-bd8f-04982dd52beb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.542853 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52a641c3-5baf-4d1d-a51f-93f372b1013e-catalog-content\") pod \"redhat-operators-9fckn\" (UID: \"52a641c3-5baf-4d1d-a51f-93f372b1013e\") " pod="openshift-marketplace/redhat-operators-9fckn" Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.563007 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93a0b3e7-600b-4a2a-bd8f-04982dd52beb-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "93a0b3e7-600b-4a2a-bd8f-04982dd52beb" (UID: "93a0b3e7-600b-4a2a-bd8f-04982dd52beb"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.563679 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93a0b3e7-600b-4a2a-bd8f-04982dd52beb-kube-api-access-tlrpd" (OuterVolumeSpecName: "kube-api-access-tlrpd") pod "93a0b3e7-600b-4a2a-bd8f-04982dd52beb" (UID: "93a0b3e7-600b-4a2a-bd8f-04982dd52beb"). InnerVolumeSpecName "kube-api-access-tlrpd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.565491 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2fgr\" (UniqueName: \"kubernetes.io/projected/52a641c3-5baf-4d1d-a51f-93f372b1013e-kube-api-access-q2fgr\") pod \"redhat-operators-9fckn\" (UID: \"52a641c3-5baf-4d1d-a51f-93f372b1013e\") " pod="openshift-marketplace/redhat-operators-9fckn" Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.597676 4733 patch_prober.go:28] interesting pod/router-default-5444994796-mq75m container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 02 15:17:51 crc kubenswrapper[4733]: [-]has-synced failed: reason withheld Feb 02 15:17:51 crc kubenswrapper[4733]: [+]process-running ok Feb 02 15:17:51 crc kubenswrapper[4733]: healthz check failed Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.597740 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mq75m" podUID="02048841-10f9-4ad2-aa2d-c2be2466a950" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.643658 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tlrpd\" (UniqueName: \"kubernetes.io/projected/93a0b3e7-600b-4a2a-bd8f-04982dd52beb-kube-api-access-tlrpd\") on node \"crc\" DevicePath \"\"" Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.643692 4733 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/93a0b3e7-600b-4a2a-bd8f-04982dd52beb-config-volume\") on node \"crc\" DevicePath \"\"" Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.643703 4733 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/93a0b3e7-600b-4a2a-bd8f-04982dd52beb-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.659677 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9fckn" Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.693724 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-frkzg"] Feb 02 15:17:51 crc kubenswrapper[4733]: E0202 15:17:51.694041 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93a0b3e7-600b-4a2a-bd8f-04982dd52beb" containerName="collect-profiles" Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.694062 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="93a0b3e7-600b-4a2a-bd8f-04982dd52beb" containerName="collect-profiles" Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.694271 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="93a0b3e7-600b-4a2a-bd8f-04982dd52beb" containerName="collect-profiles" Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.695382 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-frkzg" Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.704990 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-frkzg"] Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.751434 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 02 15:17:51 crc kubenswrapper[4733]: W0202 15:17:51.758245 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pode92fa8c6_7e1d_4185_896a_e582d573c97d.slice/crio-5ef6c9aae22039f112cdf5ada71128c93eac5175929af3a468c85f06b60fea8e WatchSource:0}: Error finding container 5ef6c9aae22039f112cdf5ada71128c93eac5175929af3a468c85f06b60fea8e: Status 404 returned error can't find the container with id 5ef6c9aae22039f112cdf5ada71128c93eac5175929af3a468c85f06b60fea8e Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.846770 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwbjv\" (UniqueName: \"kubernetes.io/projected/13882fb3-536e-4d85-ba06-4b7610a8d254-kube-api-access-nwbjv\") pod \"redhat-operators-frkzg\" (UID: \"13882fb3-536e-4d85-ba06-4b7610a8d254\") " pod="openshift-marketplace/redhat-operators-frkzg" Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.846812 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13882fb3-536e-4d85-ba06-4b7610a8d254-catalog-content\") pod \"redhat-operators-frkzg\" (UID: \"13882fb3-536e-4d85-ba06-4b7610a8d254\") " pod="openshift-marketplace/redhat-operators-frkzg" Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.846848 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13882fb3-536e-4d85-ba06-4b7610a8d254-utilities\") pod \"redhat-operators-frkzg\" (UID: \"13882fb3-536e-4d85-ba06-4b7610a8d254\") " pod="openshift-marketplace/redhat-operators-frkzg" Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.948021 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13882fb3-536e-4d85-ba06-4b7610a8d254-catalog-content\") pod \"redhat-operators-frkzg\" (UID: \"13882fb3-536e-4d85-ba06-4b7610a8d254\") " pod="openshift-marketplace/redhat-operators-frkzg" Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.948635 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13882fb3-536e-4d85-ba06-4b7610a8d254-utilities\") pod \"redhat-operators-frkzg\" (UID: \"13882fb3-536e-4d85-ba06-4b7610a8d254\") " pod="openshift-marketplace/redhat-operators-frkzg" Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.948701 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwbjv\" (UniqueName: \"kubernetes.io/projected/13882fb3-536e-4d85-ba06-4b7610a8d254-kube-api-access-nwbjv\") pod \"redhat-operators-frkzg\" (UID: \"13882fb3-536e-4d85-ba06-4b7610a8d254\") " pod="openshift-marketplace/redhat-operators-frkzg" Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.949128 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13882fb3-536e-4d85-ba06-4b7610a8d254-catalog-content\") pod \"redhat-operators-frkzg\" (UID: \"13882fb3-536e-4d85-ba06-4b7610a8d254\") " pod="openshift-marketplace/redhat-operators-frkzg" Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.949272 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13882fb3-536e-4d85-ba06-4b7610a8d254-utilities\") pod \"redhat-operators-frkzg\" (UID: \"13882fb3-536e-4d85-ba06-4b7610a8d254\") " pod="openshift-marketplace/redhat-operators-frkzg" Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.963855 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9fckn"] Feb 02 15:17:51 crc kubenswrapper[4733]: I0202 15:17:51.967963 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwbjv\" (UniqueName: \"kubernetes.io/projected/13882fb3-536e-4d85-ba06-4b7610a8d254-kube-api-access-nwbjv\") pod \"redhat-operators-frkzg\" (UID: \"13882fb3-536e-4d85-ba06-4b7610a8d254\") " pod="openshift-marketplace/redhat-operators-frkzg" Feb 02 15:17:51 crc kubenswrapper[4733]: W0202 15:17:51.978087 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod52a641c3_5baf_4d1d_a51f_93f372b1013e.slice/crio-000767d15de76401bdc4d99e1b55b5ee2cbb57e6b2c1ffb4177b58524bd5a939 WatchSource:0}: Error finding container 000767d15de76401bdc4d99e1b55b5ee2cbb57e6b2c1ffb4177b58524bd5a939: Status 404 returned error can't find the container with id 000767d15de76401bdc4d99e1b55b5ee2cbb57e6b2c1ffb4177b58524bd5a939 Feb 02 15:17:52 crc kubenswrapper[4733]: I0202 15:17:52.018181 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-frkzg" Feb 02 15:17:52 crc kubenswrapper[4733]: I0202 15:17:52.096514 4733 generic.go:334] "Generic (PLEG): container finished" podID="316793cf-5542-4af4-b6ca-ad207ecc5009" containerID="d81c7c0de1b4b673862e5550c36f78c03a1bf16814b36eee87b777da041c41a9" exitCode=0 Feb 02 15:17:52 crc kubenswrapper[4733]: I0202 15:17:52.096586 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-76wkk" event={"ID":"316793cf-5542-4af4-b6ca-ad207ecc5009","Type":"ContainerDied","Data":"d81c7c0de1b4b673862e5550c36f78c03a1bf16814b36eee87b777da041c41a9"} Feb 02 15:17:52 crc kubenswrapper[4733]: I0202 15:17:52.096612 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-76wkk" event={"ID":"316793cf-5542-4af4-b6ca-ad207ecc5009","Type":"ContainerStarted","Data":"a821ee12ee6448197d2deaf68c45ff5902304e01e0569e4a56d06a469dd005d6"} Feb 02 15:17:52 crc kubenswrapper[4733]: I0202 15:17:52.143566 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9fckn" event={"ID":"52a641c3-5baf-4d1d-a51f-93f372b1013e","Type":"ContainerStarted","Data":"000767d15de76401bdc4d99e1b55b5ee2cbb57e6b2c1ffb4177b58524bd5a939"} Feb 02 15:17:52 crc kubenswrapper[4733]: I0202 15:17:52.149055 4733 generic.go:334] "Generic (PLEG): container finished" podID="fe1172c1-0bf0-475a-b590-24c573598a08" containerID="031c19828973b7e3003deccc0f6326036019066a35683f24f1f5c3dca65664dc" exitCode=0 Feb 02 15:17:52 crc kubenswrapper[4733]: I0202 15:17:52.149140 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"fe1172c1-0bf0-475a-b590-24c573598a08","Type":"ContainerDied","Data":"031c19828973b7e3003deccc0f6326036019066a35683f24f1f5c3dca65664dc"} Feb 02 15:17:52 crc kubenswrapper[4733]: I0202 15:17:52.154017 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e92fa8c6-7e1d-4185-896a-e582d573c97d","Type":"ContainerStarted","Data":"5ef6c9aae22039f112cdf5ada71128c93eac5175929af3a468c85f06b60fea8e"} Feb 02 15:17:52 crc kubenswrapper[4733]: I0202 15:17:52.163081 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500755-fjlmd" Feb 02 15:17:52 crc kubenswrapper[4733]: I0202 15:17:52.163472 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500755-fjlmd" event={"ID":"93a0b3e7-600b-4a2a-bd8f-04982dd52beb","Type":"ContainerDied","Data":"836331d602ddd59bb84c392be4cdaf23827e96a98824ab7b73bf3b09309452e5"} Feb 02 15:17:52 crc kubenswrapper[4733]: I0202 15:17:52.163505 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="836331d602ddd59bb84c392be4cdaf23827e96a98824ab7b73bf3b09309452e5" Feb 02 15:17:52 crc kubenswrapper[4733]: I0202 15:17:52.437224 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-frkzg"] Feb 02 15:17:52 crc kubenswrapper[4733]: I0202 15:17:52.599731 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-mq75m" Feb 02 15:17:52 crc kubenswrapper[4733]: I0202 15:17:52.622411 4733 patch_prober.go:28] interesting pod/router-default-5444994796-mq75m container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 02 15:17:52 crc kubenswrapper[4733]: [+]has-synced ok Feb 02 15:17:52 crc kubenswrapper[4733]: [+]process-running ok Feb 02 15:17:52 crc kubenswrapper[4733]: healthz check failed Feb 02 15:17:52 crc kubenswrapper[4733]: I0202 15:17:52.622722 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mq75m" podUID="02048841-10f9-4ad2-aa2d-c2be2466a950" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 02 15:17:52 crc kubenswrapper[4733]: I0202 15:17:52.941874 4733 patch_prober.go:28] interesting pod/downloads-7954f5f757-wvgmn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Feb 02 15:17:52 crc kubenswrapper[4733]: I0202 15:17:52.941925 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-wvgmn" podUID="2bbf40f3-3007-4478-a1b5-09774f25b2f7" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" Feb 02 15:17:52 crc kubenswrapper[4733]: I0202 15:17:52.942224 4733 patch_prober.go:28] interesting pod/downloads-7954f5f757-wvgmn container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Feb 02 15:17:52 crc kubenswrapper[4733]: I0202 15:17:52.942240 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-wvgmn" podUID="2bbf40f3-3007-4478-a1b5-09774f25b2f7" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" Feb 02 15:17:53 crc kubenswrapper[4733]: I0202 15:17:53.199040 4733 generic.go:334] "Generic (PLEG): container finished" podID="52a641c3-5baf-4d1d-a51f-93f372b1013e" containerID="b598793b0ec7ec274a338dccfac6c1b7966f959c1eb4170470db97f776a45622" exitCode=0 Feb 02 15:17:53 crc kubenswrapper[4733]: I0202 15:17:53.199130 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9fckn" event={"ID":"52a641c3-5baf-4d1d-a51f-93f372b1013e","Type":"ContainerDied","Data":"b598793b0ec7ec274a338dccfac6c1b7966f959c1eb4170470db97f776a45622"} Feb 02 15:17:53 crc kubenswrapper[4733]: I0202 15:17:53.203232 4733 generic.go:334] "Generic (PLEG): container finished" podID="e92fa8c6-7e1d-4185-896a-e582d573c97d" containerID="cbef755c1dd7c882e089ced89b830e14ffe2886b32499078e8a295a9b299d03e" exitCode=0 Feb 02 15:17:53 crc kubenswrapper[4733]: I0202 15:17:53.203343 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e92fa8c6-7e1d-4185-896a-e582d573c97d","Type":"ContainerDied","Data":"cbef755c1dd7c882e089ced89b830e14ffe2886b32499078e8a295a9b299d03e"} Feb 02 15:17:53 crc kubenswrapper[4733]: I0202 15:17:53.208327 4733 generic.go:334] "Generic (PLEG): container finished" podID="13882fb3-536e-4d85-ba06-4b7610a8d254" containerID="e055e0e861fc98fa5a4e9c5017041d1c15e921dff15f186e2cef0874301d7ce1" exitCode=0 Feb 02 15:17:53 crc kubenswrapper[4733]: I0202 15:17:53.208989 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frkzg" event={"ID":"13882fb3-536e-4d85-ba06-4b7610a8d254","Type":"ContainerDied","Data":"e055e0e861fc98fa5a4e9c5017041d1c15e921dff15f186e2cef0874301d7ce1"} Feb 02 15:17:53 crc kubenswrapper[4733]: I0202 15:17:53.209017 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frkzg" event={"ID":"13882fb3-536e-4d85-ba06-4b7610a8d254","Type":"ContainerStarted","Data":"aa6f2970f791eff9ab7f627c41cebdb0149aa9cc926f45e3bcf71bb62e0bc9f0"} Feb 02 15:17:53 crc kubenswrapper[4733]: I0202 15:17:53.500178 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 02 15:17:53 crc kubenswrapper[4733]: I0202 15:17:53.578761 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fe1172c1-0bf0-475a-b590-24c573598a08-kubelet-dir\") pod \"fe1172c1-0bf0-475a-b590-24c573598a08\" (UID: \"fe1172c1-0bf0-475a-b590-24c573598a08\") " Feb 02 15:17:53 crc kubenswrapper[4733]: I0202 15:17:53.578897 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fe1172c1-0bf0-475a-b590-24c573598a08-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "fe1172c1-0bf0-475a-b590-24c573598a08" (UID: "fe1172c1-0bf0-475a-b590-24c573598a08"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 15:17:53 crc kubenswrapper[4733]: I0202 15:17:53.579387 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fe1172c1-0bf0-475a-b590-24c573598a08-kube-api-access\") pod \"fe1172c1-0bf0-475a-b590-24c573598a08\" (UID: \"fe1172c1-0bf0-475a-b590-24c573598a08\") " Feb 02 15:17:53 crc kubenswrapper[4733]: I0202 15:17:53.579683 4733 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fe1172c1-0bf0-475a-b590-24c573598a08-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 02 15:17:53 crc kubenswrapper[4733]: I0202 15:17:53.587312 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe1172c1-0bf0-475a-b590-24c573598a08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "fe1172c1-0bf0-475a-b590-24c573598a08" (UID: "fe1172c1-0bf0-475a-b590-24c573598a08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:17:53 crc kubenswrapper[4733]: I0202 15:17:53.601437 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-mq75m" Feb 02 15:17:53 crc kubenswrapper[4733]: I0202 15:17:53.604577 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-mq75m" Feb 02 15:17:53 crc kubenswrapper[4733]: I0202 15:17:53.680625 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fe1172c1-0bf0-475a-b590-24c573598a08-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 02 15:17:54 crc kubenswrapper[4733]: I0202 15:17:54.124073 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-4bdpn" Feb 02 15:17:54 crc kubenswrapper[4733]: I0202 15:17:54.224481 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 02 15:17:54 crc kubenswrapper[4733]: I0202 15:17:54.224941 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"fe1172c1-0bf0-475a-b590-24c573598a08","Type":"ContainerDied","Data":"b382338ee08cdcaeb010e6d398520e352586f8d8b5976bdc02a49a18ae5748a6"} Feb 02 15:17:54 crc kubenswrapper[4733]: I0202 15:17:54.224981 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b382338ee08cdcaeb010e6d398520e352586f8d8b5976bdc02a49a18ae5748a6" Feb 02 15:18:00 crc kubenswrapper[4733]: I0202 15:18:00.421557 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-f8l78" Feb 02 15:18:00 crc kubenswrapper[4733]: I0202 15:18:00.426374 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-f8l78" Feb 02 15:18:02 crc kubenswrapper[4733]: I0202 15:18:02.950586 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-wvgmn" Feb 02 15:18:03 crc kubenswrapper[4733]: I0202 15:18:03.321240 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 02 15:18:03 crc kubenswrapper[4733]: I0202 15:18:03.324059 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e92fa8c6-7e1d-4185-896a-e582d573c97d","Type":"ContainerDied","Data":"5ef6c9aae22039f112cdf5ada71128c93eac5175929af3a468c85f06b60fea8e"} Feb 02 15:18:03 crc kubenswrapper[4733]: I0202 15:18:03.324093 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5ef6c9aae22039f112cdf5ada71128c93eac5175929af3a468c85f06b60fea8e" Feb 02 15:18:03 crc kubenswrapper[4733]: I0202 15:18:03.324219 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 02 15:18:03 crc kubenswrapper[4733]: I0202 15:18:03.437494 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e92fa8c6-7e1d-4185-896a-e582d573c97d-kube-api-access\") pod \"e92fa8c6-7e1d-4185-896a-e582d573c97d\" (UID: \"e92fa8c6-7e1d-4185-896a-e582d573c97d\") " Feb 02 15:18:03 crc kubenswrapper[4733]: I0202 15:18:03.437581 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e92fa8c6-7e1d-4185-896a-e582d573c97d-kubelet-dir\") pod \"e92fa8c6-7e1d-4185-896a-e582d573c97d\" (UID: \"e92fa8c6-7e1d-4185-896a-e582d573c97d\") " Feb 02 15:18:03 crc kubenswrapper[4733]: I0202 15:18:03.437625 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e92fa8c6-7e1d-4185-896a-e582d573c97d-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "e92fa8c6-7e1d-4185-896a-e582d573c97d" (UID: "e92fa8c6-7e1d-4185-896a-e582d573c97d"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 15:18:03 crc kubenswrapper[4733]: I0202 15:18:03.437907 4733 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e92fa8c6-7e1d-4185-896a-e582d573c97d-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 02 15:18:03 crc kubenswrapper[4733]: I0202 15:18:03.452722 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e92fa8c6-7e1d-4185-896a-e582d573c97d-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e92fa8c6-7e1d-4185-896a-e582d573c97d" (UID: "e92fa8c6-7e1d-4185-896a-e582d573c97d"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:18:03 crc kubenswrapper[4733]: I0202 15:18:03.539440 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e92fa8c6-7e1d-4185-896a-e582d573c97d-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 02 15:18:04 crc kubenswrapper[4733]: I0202 15:18:04.987952 4733 patch_prober.go:28] interesting pod/machine-config-daemon-zgsv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 15:18:04 crc kubenswrapper[4733]: I0202 15:18:04.988261 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 15:18:08 crc kubenswrapper[4733]: I0202 15:18:08.581005 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:18:22 crc kubenswrapper[4733]: I0202 15:18:22.311046 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xhg5n" Feb 02 15:18:22 crc kubenswrapper[4733]: I0202 15:18:22.440859 4733 generic.go:334] "Generic (PLEG): container finished" podID="d116a9fc-670f-4a77-b9ce-d6903d2c3d24" containerID="ef0f3f1bc1e12fba4768904f745075561d388640cd5e0889dc61fe8c1376ae26" exitCode=0 Feb 02 15:18:22 crc kubenswrapper[4733]: I0202 15:18:22.440943 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wmd4l" event={"ID":"d116a9fc-670f-4a77-b9ce-d6903d2c3d24","Type":"ContainerDied","Data":"ef0f3f1bc1e12fba4768904f745075561d388640cd5e0889dc61fe8c1376ae26"} Feb 02 15:18:22 crc kubenswrapper[4733]: I0202 15:18:22.443401 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9fckn" event={"ID":"52a641c3-5baf-4d1d-a51f-93f372b1013e","Type":"ContainerStarted","Data":"444ff31d550de20fdebc035f8168dc30ce053ded9043e88036b8f6cb1f3adde4"} Feb 02 15:18:22 crc kubenswrapper[4733]: I0202 15:18:22.449026 4733 generic.go:334] "Generic (PLEG): container finished" podID="7b6ed8e1-1992-4d83-954a-552aa8e7acb3" containerID="b217ce646f132f2a839baaa49cbb8544b12840ec5f9d6e1bf39efc02b87cec16" exitCode=0 Feb 02 15:18:22 crc kubenswrapper[4733]: I0202 15:18:22.449078 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v5t9s" event={"ID":"7b6ed8e1-1992-4d83-954a-552aa8e7acb3","Type":"ContainerDied","Data":"b217ce646f132f2a839baaa49cbb8544b12840ec5f9d6e1bf39efc02b87cec16"} Feb 02 15:18:22 crc kubenswrapper[4733]: I0202 15:18:22.451635 4733 generic.go:334] "Generic (PLEG): container finished" podID="d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2" containerID="0cdc71ed22d00abcd62c5b8053ec06dea21eb456bf11c35da7d64cdf192f1f83" exitCode=0 Feb 02 15:18:22 crc kubenswrapper[4733]: I0202 15:18:22.451804 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7gm8t" event={"ID":"d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2","Type":"ContainerDied","Data":"0cdc71ed22d00abcd62c5b8053ec06dea21eb456bf11c35da7d64cdf192f1f83"} Feb 02 15:18:22 crc kubenswrapper[4733]: I0202 15:18:22.453344 4733 generic.go:334] "Generic (PLEG): container finished" podID="5f14a53f-b3e0-46ab-af53-2dfb4041498d" containerID="86bea7dcb5d2d734c5c76e7cc515dd97fb0b66c77788d02adee70f2855dd3727" exitCode=0 Feb 02 15:18:22 crc kubenswrapper[4733]: I0202 15:18:22.453404 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m6bvv" event={"ID":"5f14a53f-b3e0-46ab-af53-2dfb4041498d","Type":"ContainerDied","Data":"86bea7dcb5d2d734c5c76e7cc515dd97fb0b66c77788d02adee70f2855dd3727"} Feb 02 15:18:22 crc kubenswrapper[4733]: I0202 15:18:22.456734 4733 generic.go:334] "Generic (PLEG): container finished" podID="316793cf-5542-4af4-b6ca-ad207ecc5009" containerID="ce5e08f104b288ab3d5d9415126837e1d6fbf5e15520f2881b6c312c3bc56e3f" exitCode=0 Feb 02 15:18:22 crc kubenswrapper[4733]: I0202 15:18:22.456780 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-76wkk" event={"ID":"316793cf-5542-4af4-b6ca-ad207ecc5009","Type":"ContainerDied","Data":"ce5e08f104b288ab3d5d9415126837e1d6fbf5e15520f2881b6c312c3bc56e3f"} Feb 02 15:18:22 crc kubenswrapper[4733]: I0202 15:18:22.458869 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frkzg" event={"ID":"13882fb3-536e-4d85-ba06-4b7610a8d254","Type":"ContainerStarted","Data":"0afb3f4a8eaed5fa5ab6de21c6d633669c87ece98941e3071f58f97663f98f85"} Feb 02 15:18:22 crc kubenswrapper[4733]: I0202 15:18:22.463420 4733 generic.go:334] "Generic (PLEG): container finished" podID="4f03c333-e449-4a18-8efb-f8834911a599" containerID="4ebe46184910bfd2a590af1455b2c9e53b259a81917e5aebabb1c387b30f8898" exitCode=0 Feb 02 15:18:22 crc kubenswrapper[4733]: I0202 15:18:22.463458 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jk5b2" event={"ID":"4f03c333-e449-4a18-8efb-f8834911a599","Type":"ContainerDied","Data":"4ebe46184910bfd2a590af1455b2c9e53b259a81917e5aebabb1c387b30f8898"} Feb 02 15:18:23 crc kubenswrapper[4733]: I0202 15:18:23.473935 4733 generic.go:334] "Generic (PLEG): container finished" podID="13882fb3-536e-4d85-ba06-4b7610a8d254" containerID="0afb3f4a8eaed5fa5ab6de21c6d633669c87ece98941e3071f58f97663f98f85" exitCode=0 Feb 02 15:18:23 crc kubenswrapper[4733]: I0202 15:18:23.474220 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frkzg" event={"ID":"13882fb3-536e-4d85-ba06-4b7610a8d254","Type":"ContainerDied","Data":"0afb3f4a8eaed5fa5ab6de21c6d633669c87ece98941e3071f58f97663f98f85"} Feb 02 15:18:23 crc kubenswrapper[4733]: I0202 15:18:23.482846 4733 generic.go:334] "Generic (PLEG): container finished" podID="52a641c3-5baf-4d1d-a51f-93f372b1013e" containerID="444ff31d550de20fdebc035f8168dc30ce053ded9043e88036b8f6cb1f3adde4" exitCode=0 Feb 02 15:18:23 crc kubenswrapper[4733]: I0202 15:18:23.482920 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9fckn" event={"ID":"52a641c3-5baf-4d1d-a51f-93f372b1013e","Type":"ContainerDied","Data":"444ff31d550de20fdebc035f8168dc30ce053ded9043e88036b8f6cb1f3adde4"} Feb 02 15:18:25 crc kubenswrapper[4733]: I0202 15:18:25.337154 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 02 15:18:25 crc kubenswrapper[4733]: I0202 15:18:25.713757 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 02 15:18:25 crc kubenswrapper[4733]: E0202 15:18:25.714076 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e92fa8c6-7e1d-4185-896a-e582d573c97d" containerName="pruner" Feb 02 15:18:25 crc kubenswrapper[4733]: I0202 15:18:25.714092 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="e92fa8c6-7e1d-4185-896a-e582d573c97d" containerName="pruner" Feb 02 15:18:25 crc kubenswrapper[4733]: E0202 15:18:25.714111 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe1172c1-0bf0-475a-b590-24c573598a08" containerName="pruner" Feb 02 15:18:25 crc kubenswrapper[4733]: I0202 15:18:25.714119 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe1172c1-0bf0-475a-b590-24c573598a08" containerName="pruner" Feb 02 15:18:25 crc kubenswrapper[4733]: I0202 15:18:25.714267 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="e92fa8c6-7e1d-4185-896a-e582d573c97d" containerName="pruner" Feb 02 15:18:25 crc kubenswrapper[4733]: I0202 15:18:25.714282 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe1172c1-0bf0-475a-b590-24c573598a08" containerName="pruner" Feb 02 15:18:25 crc kubenswrapper[4733]: I0202 15:18:25.714724 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 02 15:18:25 crc kubenswrapper[4733]: I0202 15:18:25.718033 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 02 15:18:25 crc kubenswrapper[4733]: I0202 15:18:25.718208 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 02 15:18:25 crc kubenswrapper[4733]: I0202 15:18:25.723384 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 02 15:18:25 crc kubenswrapper[4733]: I0202 15:18:25.745083 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/929b6d17-4fed-4440-8e42-df3655c01be7-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"929b6d17-4fed-4440-8e42-df3655c01be7\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 02 15:18:25 crc kubenswrapper[4733]: I0202 15:18:25.745272 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/929b6d17-4fed-4440-8e42-df3655c01be7-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"929b6d17-4fed-4440-8e42-df3655c01be7\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 02 15:18:25 crc kubenswrapper[4733]: I0202 15:18:25.847017 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/929b6d17-4fed-4440-8e42-df3655c01be7-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"929b6d17-4fed-4440-8e42-df3655c01be7\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 02 15:18:25 crc kubenswrapper[4733]: I0202 15:18:25.847199 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/929b6d17-4fed-4440-8e42-df3655c01be7-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"929b6d17-4fed-4440-8e42-df3655c01be7\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 02 15:18:25 crc kubenswrapper[4733]: I0202 15:18:25.847252 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/929b6d17-4fed-4440-8e42-df3655c01be7-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"929b6d17-4fed-4440-8e42-df3655c01be7\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 02 15:18:25 crc kubenswrapper[4733]: I0202 15:18:25.879193 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/929b6d17-4fed-4440-8e42-df3655c01be7-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"929b6d17-4fed-4440-8e42-df3655c01be7\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 02 15:18:26 crc kubenswrapper[4733]: I0202 15:18:26.062675 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 02 15:18:26 crc kubenswrapper[4733]: I0202 15:18:26.511919 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wmd4l" event={"ID":"d116a9fc-670f-4a77-b9ce-d6903d2c3d24","Type":"ContainerStarted","Data":"ccc1c1d72e9362808fffde1f90fc4cb5a536eac15e82ed2269d32b042a72f15b"} Feb 02 15:18:26 crc kubenswrapper[4733]: I0202 15:18:26.535264 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 02 15:18:26 crc kubenswrapper[4733]: W0202 15:18:26.571986 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod929b6d17_4fed_4440_8e42_df3655c01be7.slice/crio-6d3f345dcb73096e2dd11f9b3699391615e5154b5dad2dc7546e5c7777cc375c WatchSource:0}: Error finding container 6d3f345dcb73096e2dd11f9b3699391615e5154b5dad2dc7546e5c7777cc375c: Status 404 returned error can't find the container with id 6d3f345dcb73096e2dd11f9b3699391615e5154b5dad2dc7546e5c7777cc375c Feb 02 15:18:27 crc kubenswrapper[4733]: I0202 15:18:27.519175 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"929b6d17-4fed-4440-8e42-df3655c01be7","Type":"ContainerStarted","Data":"0864ad36e0a8f3a459a275f57cc474cb569a2794ef79325e51aa98b9fae299ff"} Feb 02 15:18:27 crc kubenswrapper[4733]: I0202 15:18:27.519480 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"929b6d17-4fed-4440-8e42-df3655c01be7","Type":"ContainerStarted","Data":"6d3f345dcb73096e2dd11f9b3699391615e5154b5dad2dc7546e5c7777cc375c"} Feb 02 15:18:27 crc kubenswrapper[4733]: I0202 15:18:27.546405 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-wmd4l" podStartSLOduration=3.986846652 podStartE2EDuration="39.546377836s" podCreationTimestamp="2026-02-02 15:17:48 +0000 UTC" firstStartedPulling="2026-02-02 15:17:50.02254804 +0000 UTC m=+153.474009388" lastFinishedPulling="2026-02-02 15:18:25.582079194 +0000 UTC m=+189.033540572" observedRunningTime="2026-02-02 15:18:27.538845785 +0000 UTC m=+190.990307163" watchObservedRunningTime="2026-02-02 15:18:27.546377836 +0000 UTC m=+190.997839204" Feb 02 15:18:28 crc kubenswrapper[4733]: I0202 15:18:28.567667 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=3.5676356670000002 podStartE2EDuration="3.567635667s" podCreationTimestamp="2026-02-02 15:18:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:18:28.553247353 +0000 UTC m=+192.004708771" watchObservedRunningTime="2026-02-02 15:18:28.567635667 +0000 UTC m=+192.019097055" Feb 02 15:18:29 crc kubenswrapper[4733]: I0202 15:18:29.012313 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-wmd4l" Feb 02 15:18:29 crc kubenswrapper[4733]: I0202 15:18:29.012641 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-wmd4l" Feb 02 15:18:29 crc kubenswrapper[4733]: I0202 15:18:29.533808 4733 generic.go:334] "Generic (PLEG): container finished" podID="929b6d17-4fed-4440-8e42-df3655c01be7" containerID="0864ad36e0a8f3a459a275f57cc474cb569a2794ef79325e51aa98b9fae299ff" exitCode=0 Feb 02 15:18:29 crc kubenswrapper[4733]: I0202 15:18:29.533870 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"929b6d17-4fed-4440-8e42-df3655c01be7","Type":"ContainerDied","Data":"0864ad36e0a8f3a459a275f57cc474cb569a2794ef79325e51aa98b9fae299ff"} Feb 02 15:18:30 crc kubenswrapper[4733]: I0202 15:18:30.978329 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-wmd4l" podUID="d116a9fc-670f-4a77-b9ce-d6903d2c3d24" containerName="registry-server" probeResult="failure" output=< Feb 02 15:18:30 crc kubenswrapper[4733]: timeout: failed to connect service ":50051" within 1s Feb 02 15:18:30 crc kubenswrapper[4733]: > Feb 02 15:18:31 crc kubenswrapper[4733]: I0202 15:18:31.381976 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 02 15:18:31 crc kubenswrapper[4733]: I0202 15:18:31.429018 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/929b6d17-4fed-4440-8e42-df3655c01be7-kube-api-access\") pod \"929b6d17-4fed-4440-8e42-df3655c01be7\" (UID: \"929b6d17-4fed-4440-8e42-df3655c01be7\") " Feb 02 15:18:31 crc kubenswrapper[4733]: I0202 15:18:31.429128 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/929b6d17-4fed-4440-8e42-df3655c01be7-kubelet-dir\") pod \"929b6d17-4fed-4440-8e42-df3655c01be7\" (UID: \"929b6d17-4fed-4440-8e42-df3655c01be7\") " Feb 02 15:18:31 crc kubenswrapper[4733]: I0202 15:18:31.429250 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/929b6d17-4fed-4440-8e42-df3655c01be7-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "929b6d17-4fed-4440-8e42-df3655c01be7" (UID: "929b6d17-4fed-4440-8e42-df3655c01be7"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 15:18:31 crc kubenswrapper[4733]: I0202 15:18:31.429593 4733 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/929b6d17-4fed-4440-8e42-df3655c01be7-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 02 15:18:31 crc kubenswrapper[4733]: I0202 15:18:31.435446 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/929b6d17-4fed-4440-8e42-df3655c01be7-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "929b6d17-4fed-4440-8e42-df3655c01be7" (UID: "929b6d17-4fed-4440-8e42-df3655c01be7"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:18:31 crc kubenswrapper[4733]: I0202 15:18:31.530950 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/929b6d17-4fed-4440-8e42-df3655c01be7-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 02 15:18:31 crc kubenswrapper[4733]: I0202 15:18:31.549498 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"929b6d17-4fed-4440-8e42-df3655c01be7","Type":"ContainerDied","Data":"6d3f345dcb73096e2dd11f9b3699391615e5154b5dad2dc7546e5c7777cc375c"} Feb 02 15:18:31 crc kubenswrapper[4733]: I0202 15:18:31.549642 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d3f345dcb73096e2dd11f9b3699391615e5154b5dad2dc7546e5c7777cc375c" Feb 02 15:18:31 crc kubenswrapper[4733]: I0202 15:18:31.549606 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 02 15:18:31 crc kubenswrapper[4733]: I0202 15:18:31.895676 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 02 15:18:31 crc kubenswrapper[4733]: E0202 15:18:31.895937 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="929b6d17-4fed-4440-8e42-df3655c01be7" containerName="pruner" Feb 02 15:18:31 crc kubenswrapper[4733]: I0202 15:18:31.895952 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="929b6d17-4fed-4440-8e42-df3655c01be7" containerName="pruner" Feb 02 15:18:31 crc kubenswrapper[4733]: I0202 15:18:31.896089 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="929b6d17-4fed-4440-8e42-df3655c01be7" containerName="pruner" Feb 02 15:18:31 crc kubenswrapper[4733]: I0202 15:18:31.896567 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 02 15:18:31 crc kubenswrapper[4733]: I0202 15:18:31.905549 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 02 15:18:31 crc kubenswrapper[4733]: I0202 15:18:31.907922 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 02 15:18:31 crc kubenswrapper[4733]: I0202 15:18:31.914050 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 02 15:18:31 crc kubenswrapper[4733]: I0202 15:18:31.940222 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/272ea61e-a61b-40c8-8e4a-cdc9f3c95847-var-lock\") pod \"installer-9-crc\" (UID: \"272ea61e-a61b-40c8-8e4a-cdc9f3c95847\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 02 15:18:31 crc kubenswrapper[4733]: I0202 15:18:31.940375 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/272ea61e-a61b-40c8-8e4a-cdc9f3c95847-kubelet-dir\") pod \"installer-9-crc\" (UID: \"272ea61e-a61b-40c8-8e4a-cdc9f3c95847\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 02 15:18:31 crc kubenswrapper[4733]: I0202 15:18:31.940465 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/272ea61e-a61b-40c8-8e4a-cdc9f3c95847-kube-api-access\") pod \"installer-9-crc\" (UID: \"272ea61e-a61b-40c8-8e4a-cdc9f3c95847\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 02 15:18:32 crc kubenswrapper[4733]: I0202 15:18:32.042735 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/272ea61e-a61b-40c8-8e4a-cdc9f3c95847-kube-api-access\") pod \"installer-9-crc\" (UID: \"272ea61e-a61b-40c8-8e4a-cdc9f3c95847\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 02 15:18:32 crc kubenswrapper[4733]: I0202 15:18:32.042856 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/272ea61e-a61b-40c8-8e4a-cdc9f3c95847-var-lock\") pod \"installer-9-crc\" (UID: \"272ea61e-a61b-40c8-8e4a-cdc9f3c95847\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 02 15:18:32 crc kubenswrapper[4733]: I0202 15:18:32.042912 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/272ea61e-a61b-40c8-8e4a-cdc9f3c95847-kubelet-dir\") pod \"installer-9-crc\" (UID: \"272ea61e-a61b-40c8-8e4a-cdc9f3c95847\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 02 15:18:32 crc kubenswrapper[4733]: I0202 15:18:32.043040 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/272ea61e-a61b-40c8-8e4a-cdc9f3c95847-kubelet-dir\") pod \"installer-9-crc\" (UID: \"272ea61e-a61b-40c8-8e4a-cdc9f3c95847\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 02 15:18:32 crc kubenswrapper[4733]: I0202 15:18:32.043056 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/272ea61e-a61b-40c8-8e4a-cdc9f3c95847-var-lock\") pod \"installer-9-crc\" (UID: \"272ea61e-a61b-40c8-8e4a-cdc9f3c95847\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 02 15:18:32 crc kubenswrapper[4733]: I0202 15:18:32.070918 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/272ea61e-a61b-40c8-8e4a-cdc9f3c95847-kube-api-access\") pod \"installer-9-crc\" (UID: \"272ea61e-a61b-40c8-8e4a-cdc9f3c95847\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 02 15:18:32 crc kubenswrapper[4733]: I0202 15:18:32.232759 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 02 15:18:34 crc kubenswrapper[4733]: I0202 15:18:34.988071 4733 patch_prober.go:28] interesting pod/machine-config-daemon-zgsv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 15:18:34 crc kubenswrapper[4733]: I0202 15:18:34.988694 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 15:18:38 crc kubenswrapper[4733]: I0202 15:18:38.471679 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 02 15:18:38 crc kubenswrapper[4733]: I0202 15:18:38.600432 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"272ea61e-a61b-40c8-8e4a-cdc9f3c95847","Type":"ContainerStarted","Data":"812ca4ef6743546483882c26c88426865bf39206c0ffc4e3b25472e6f5588151"} Feb 02 15:18:39 crc kubenswrapper[4733]: I0202 15:18:39.307573 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-wmd4l" Feb 02 15:18:39 crc kubenswrapper[4733]: I0202 15:18:39.378899 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-wmd4l" Feb 02 15:18:39 crc kubenswrapper[4733]: I0202 15:18:39.550231 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wmd4l"] Feb 02 15:18:39 crc kubenswrapper[4733]: I0202 15:18:39.609759 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frkzg" event={"ID":"13882fb3-536e-4d85-ba06-4b7610a8d254","Type":"ContainerStarted","Data":"2862f3e13edc819755f766140152b4d94e06fb9ed85371e2f305d867a0443d57"} Feb 02 15:18:39 crc kubenswrapper[4733]: I0202 15:18:39.611996 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jk5b2" event={"ID":"4f03c333-e449-4a18-8efb-f8834911a599","Type":"ContainerStarted","Data":"e4a4c0a9fa82721f255c6e6572b720748b1dd22edba932206f6daaf366db2c53"} Feb 02 15:18:39 crc kubenswrapper[4733]: I0202 15:18:39.614399 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9fckn" event={"ID":"52a641c3-5baf-4d1d-a51f-93f372b1013e","Type":"ContainerStarted","Data":"bc16b500617a150f81ab1c82995e4be4e445957acddb58d55773c44c85ff85e7"} Feb 02 15:18:39 crc kubenswrapper[4733]: I0202 15:18:39.616827 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v5t9s" event={"ID":"7b6ed8e1-1992-4d83-954a-552aa8e7acb3","Type":"ContainerStarted","Data":"c9aa09ab01768bd69cd858d98490db679284ae40c82f880a58ece8b87c424d3f"} Feb 02 15:18:39 crc kubenswrapper[4733]: I0202 15:18:39.618897 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7gm8t" event={"ID":"d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2","Type":"ContainerStarted","Data":"720a4dabd902a96d042261a42019bb9ea3beee89af69a1d942baecce6063882e"} Feb 02 15:18:39 crc kubenswrapper[4733]: I0202 15:18:39.622069 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m6bvv" event={"ID":"5f14a53f-b3e0-46ab-af53-2dfb4041498d","Type":"ContainerStarted","Data":"c31ab2851ff2d0321b9b9a7e33773d9cd0cb7f4a078829841f34c98e26972eba"} Feb 02 15:18:39 crc kubenswrapper[4733]: I0202 15:18:39.625584 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-76wkk" event={"ID":"316793cf-5542-4af4-b6ca-ad207ecc5009","Type":"ContainerStarted","Data":"996fc24a0c61b46b3872ab67c673a08e2060740f6c805e2dd382dbbcf213ebf6"} Feb 02 15:18:39 crc kubenswrapper[4733]: I0202 15:18:39.627417 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"272ea61e-a61b-40c8-8e4a-cdc9f3c95847","Type":"ContainerStarted","Data":"a501edad55227183ad5dfc3a0ac0dfec399c4b05bc14706420020507a35c6599"} Feb 02 15:18:39 crc kubenswrapper[4733]: I0202 15:18:39.633495 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jk5b2" podStartSLOduration=11.038658783 podStartE2EDuration="51.633477404s" podCreationTimestamp="2026-02-02 15:17:48 +0000 UTC" firstStartedPulling="2026-02-02 15:17:50.017271289 +0000 UTC m=+153.468732647" lastFinishedPulling="2026-02-02 15:18:30.61208987 +0000 UTC m=+194.063551268" observedRunningTime="2026-02-02 15:18:39.629933201 +0000 UTC m=+203.081394569" watchObservedRunningTime="2026-02-02 15:18:39.633477404 +0000 UTC m=+203.084938772" Feb 02 15:18:39 crc kubenswrapper[4733]: I0202 15:18:39.648650 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7gm8t" podStartSLOduration=3.315790498 podStartE2EDuration="51.648634528s" podCreationTimestamp="2026-02-02 15:17:48 +0000 UTC" firstStartedPulling="2026-02-02 15:17:48.985871887 +0000 UTC m=+152.437333245" lastFinishedPulling="2026-02-02 15:18:37.318715917 +0000 UTC m=+200.770177275" observedRunningTime="2026-02-02 15:18:39.64525871 +0000 UTC m=+203.096720068" watchObservedRunningTime="2026-02-02 15:18:39.648634528 +0000 UTC m=+203.100095886" Feb 02 15:18:40 crc kubenswrapper[4733]: I0202 15:18:40.631308 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-wmd4l" podUID="d116a9fc-670f-4a77-b9ce-d6903d2c3d24" containerName="registry-server" containerID="cri-o://ccc1c1d72e9362808fffde1f90fc4cb5a536eac15e82ed2269d32b042a72f15b" gracePeriod=2 Feb 02 15:18:40 crc kubenswrapper[4733]: I0202 15:18:40.674811 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9fckn" podStartSLOduration=8.494966931 podStartE2EDuration="49.674794803s" podCreationTimestamp="2026-02-02 15:17:51 +0000 UTC" firstStartedPulling="2026-02-02 15:17:53.200942354 +0000 UTC m=+156.652403712" lastFinishedPulling="2026-02-02 15:18:34.380770216 +0000 UTC m=+197.832231584" observedRunningTime="2026-02-02 15:18:40.671504338 +0000 UTC m=+204.122965686" watchObservedRunningTime="2026-02-02 15:18:40.674794803 +0000 UTC m=+204.126256161" Feb 02 15:18:40 crc kubenswrapper[4733]: I0202 15:18:40.677379 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-frkzg" podStartSLOduration=4.875904473 podStartE2EDuration="49.677372529s" podCreationTimestamp="2026-02-02 15:17:51 +0000 UTC" firstStartedPulling="2026-02-02 15:17:53.210311925 +0000 UTC m=+156.661773283" lastFinishedPulling="2026-02-02 15:18:38.011779941 +0000 UTC m=+201.463241339" observedRunningTime="2026-02-02 15:18:40.654328191 +0000 UTC m=+204.105789549" watchObservedRunningTime="2026-02-02 15:18:40.677372529 +0000 UTC m=+204.128833887" Feb 02 15:18:40 crc kubenswrapper[4733]: I0202 15:18:40.695789 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-v5t9s" podStartSLOduration=3.7516059840000002 podStartE2EDuration="50.695770518s" podCreationTimestamp="2026-02-02 15:17:50 +0000 UTC" firstStartedPulling="2026-02-02 15:17:51.068838269 +0000 UTC m=+154.520299627" lastFinishedPulling="2026-02-02 15:18:38.013002763 +0000 UTC m=+201.464464161" observedRunningTime="2026-02-02 15:18:40.692365249 +0000 UTC m=+204.143826617" watchObservedRunningTime="2026-02-02 15:18:40.695770518 +0000 UTC m=+204.147231876" Feb 02 15:18:40 crc kubenswrapper[4733]: I0202 15:18:40.706768 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=9.706748933 podStartE2EDuration="9.706748933s" podCreationTimestamp="2026-02-02 15:18:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:18:40.705145502 +0000 UTC m=+204.156606850" watchObservedRunningTime="2026-02-02 15:18:40.706748933 +0000 UTC m=+204.158210291" Feb 02 15:18:40 crc kubenswrapper[4733]: I0202 15:18:40.747236 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-m6bvv" podStartSLOduration=6.268542157 podStartE2EDuration="52.747218296s" podCreationTimestamp="2026-02-02 15:17:48 +0000 UTC" firstStartedPulling="2026-02-02 15:17:50.009719187 +0000 UTC m=+153.461180545" lastFinishedPulling="2026-02-02 15:18:36.488395286 +0000 UTC m=+199.939856684" observedRunningTime="2026-02-02 15:18:40.745964313 +0000 UTC m=+204.197425681" watchObservedRunningTime="2026-02-02 15:18:40.747218296 +0000 UTC m=+204.198679654" Feb 02 15:18:40 crc kubenswrapper[4733]: I0202 15:18:40.747798 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-76wkk" podStartSLOduration=4.633068555 podStartE2EDuration="50.747792141s" podCreationTimestamp="2026-02-02 15:17:50 +0000 UTC" firstStartedPulling="2026-02-02 15:17:52.124837468 +0000 UTC m=+155.576298826" lastFinishedPulling="2026-02-02 15:18:38.239561014 +0000 UTC m=+201.691022412" observedRunningTime="2026-02-02 15:18:40.727702888 +0000 UTC m=+204.179164246" watchObservedRunningTime="2026-02-02 15:18:40.747792141 +0000 UTC m=+204.199253499" Feb 02 15:18:40 crc kubenswrapper[4733]: I0202 15:18:40.830559 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-76wkk" Feb 02 15:18:40 crc kubenswrapper[4733]: I0202 15:18:40.830601 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-76wkk" Feb 02 15:18:40 crc kubenswrapper[4733]: I0202 15:18:40.950178 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-4cq4j"] Feb 02 15:18:41 crc kubenswrapper[4733]: I0202 15:18:41.660071 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9fckn" Feb 02 15:18:41 crc kubenswrapper[4733]: I0202 15:18:41.660343 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9fckn" Feb 02 15:18:41 crc kubenswrapper[4733]: I0202 15:18:41.886071 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-76wkk" podUID="316793cf-5542-4af4-b6ca-ad207ecc5009" containerName="registry-server" probeResult="failure" output=< Feb 02 15:18:41 crc kubenswrapper[4733]: timeout: failed to connect service ":50051" within 1s Feb 02 15:18:41 crc kubenswrapper[4733]: > Feb 02 15:18:42 crc kubenswrapper[4733]: I0202 15:18:42.018886 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-frkzg" Feb 02 15:18:42 crc kubenswrapper[4733]: I0202 15:18:42.018934 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-frkzg" Feb 02 15:18:42 crc kubenswrapper[4733]: I0202 15:18:42.641382 4733 generic.go:334] "Generic (PLEG): container finished" podID="d116a9fc-670f-4a77-b9ce-d6903d2c3d24" containerID="ccc1c1d72e9362808fffde1f90fc4cb5a536eac15e82ed2269d32b042a72f15b" exitCode=0 Feb 02 15:18:42 crc kubenswrapper[4733]: I0202 15:18:42.641429 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wmd4l" event={"ID":"d116a9fc-670f-4a77-b9ce-d6903d2c3d24","Type":"ContainerDied","Data":"ccc1c1d72e9362808fffde1f90fc4cb5a536eac15e82ed2269d32b042a72f15b"} Feb 02 15:18:42 crc kubenswrapper[4733]: I0202 15:18:42.710804 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9fckn" podUID="52a641c3-5baf-4d1d-a51f-93f372b1013e" containerName="registry-server" probeResult="failure" output=< Feb 02 15:18:42 crc kubenswrapper[4733]: timeout: failed to connect service ":50051" within 1s Feb 02 15:18:42 crc kubenswrapper[4733]: > Feb 02 15:18:42 crc kubenswrapper[4733]: I0202 15:18:42.861349 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wmd4l" Feb 02 15:18:42 crc kubenswrapper[4733]: I0202 15:18:42.950415 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d116a9fc-670f-4a77-b9ce-d6903d2c3d24-utilities\") pod \"d116a9fc-670f-4a77-b9ce-d6903d2c3d24\" (UID: \"d116a9fc-670f-4a77-b9ce-d6903d2c3d24\") " Feb 02 15:18:42 crc kubenswrapper[4733]: I0202 15:18:42.950513 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9jwh9\" (UniqueName: \"kubernetes.io/projected/d116a9fc-670f-4a77-b9ce-d6903d2c3d24-kube-api-access-9jwh9\") pod \"d116a9fc-670f-4a77-b9ce-d6903d2c3d24\" (UID: \"d116a9fc-670f-4a77-b9ce-d6903d2c3d24\") " Feb 02 15:18:42 crc kubenswrapper[4733]: I0202 15:18:42.950570 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d116a9fc-670f-4a77-b9ce-d6903d2c3d24-catalog-content\") pod \"d116a9fc-670f-4a77-b9ce-d6903d2c3d24\" (UID: \"d116a9fc-670f-4a77-b9ce-d6903d2c3d24\") " Feb 02 15:18:42 crc kubenswrapper[4733]: I0202 15:18:42.951367 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d116a9fc-670f-4a77-b9ce-d6903d2c3d24-utilities" (OuterVolumeSpecName: "utilities") pod "d116a9fc-670f-4a77-b9ce-d6903d2c3d24" (UID: "d116a9fc-670f-4a77-b9ce-d6903d2c3d24"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:18:42 crc kubenswrapper[4733]: I0202 15:18:42.957353 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d116a9fc-670f-4a77-b9ce-d6903d2c3d24-kube-api-access-9jwh9" (OuterVolumeSpecName: "kube-api-access-9jwh9") pod "d116a9fc-670f-4a77-b9ce-d6903d2c3d24" (UID: "d116a9fc-670f-4a77-b9ce-d6903d2c3d24"). InnerVolumeSpecName "kube-api-access-9jwh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:18:43 crc kubenswrapper[4733]: I0202 15:18:43.003000 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d116a9fc-670f-4a77-b9ce-d6903d2c3d24-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d116a9fc-670f-4a77-b9ce-d6903d2c3d24" (UID: "d116a9fc-670f-4a77-b9ce-d6903d2c3d24"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:18:43 crc kubenswrapper[4733]: I0202 15:18:43.051689 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d116a9fc-670f-4a77-b9ce-d6903d2c3d24-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 15:18:43 crc kubenswrapper[4733]: I0202 15:18:43.051734 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9jwh9\" (UniqueName: \"kubernetes.io/projected/d116a9fc-670f-4a77-b9ce-d6903d2c3d24-kube-api-access-9jwh9\") on node \"crc\" DevicePath \"\"" Feb 02 15:18:43 crc kubenswrapper[4733]: I0202 15:18:43.051750 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d116a9fc-670f-4a77-b9ce-d6903d2c3d24-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 15:18:43 crc kubenswrapper[4733]: I0202 15:18:43.059199 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-frkzg" podUID="13882fb3-536e-4d85-ba06-4b7610a8d254" containerName="registry-server" probeResult="failure" output=< Feb 02 15:18:43 crc kubenswrapper[4733]: timeout: failed to connect service ":50051" within 1s Feb 02 15:18:43 crc kubenswrapper[4733]: > Feb 02 15:18:43 crc kubenswrapper[4733]: I0202 15:18:43.648604 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wmd4l" event={"ID":"d116a9fc-670f-4a77-b9ce-d6903d2c3d24","Type":"ContainerDied","Data":"9e692b73f4ff219635cbbe4910a8408f697662bc9f48fe6ca665875ae99b1e49"} Feb 02 15:18:43 crc kubenswrapper[4733]: I0202 15:18:43.648663 4733 scope.go:117] "RemoveContainer" containerID="ccc1c1d72e9362808fffde1f90fc4cb5a536eac15e82ed2269d32b042a72f15b" Feb 02 15:18:43 crc kubenswrapper[4733]: I0202 15:18:43.648670 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wmd4l" Feb 02 15:18:43 crc kubenswrapper[4733]: I0202 15:18:43.670853 4733 scope.go:117] "RemoveContainer" containerID="ef0f3f1bc1e12fba4768904f745075561d388640cd5e0889dc61fe8c1376ae26" Feb 02 15:18:43 crc kubenswrapper[4733]: I0202 15:18:43.675197 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wmd4l"] Feb 02 15:18:43 crc kubenswrapper[4733]: I0202 15:18:43.678020 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-wmd4l"] Feb 02 15:18:43 crc kubenswrapper[4733]: I0202 15:18:43.698858 4733 scope.go:117] "RemoveContainer" containerID="c1d608ae3eb57339ae0b401ec54724b0ec2601dc46b6e145a1a425f431eab741" Feb 02 15:18:45 crc kubenswrapper[4733]: I0202 15:18:45.265949 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d116a9fc-670f-4a77-b9ce-d6903d2c3d24" path="/var/lib/kubelet/pods/d116a9fc-670f-4a77-b9ce-d6903d2c3d24/volumes" Feb 02 15:18:48 crc kubenswrapper[4733]: I0202 15:18:48.423925 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7gm8t" Feb 02 15:18:48 crc kubenswrapper[4733]: I0202 15:18:48.424378 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7gm8t" Feb 02 15:18:48 crc kubenswrapper[4733]: I0202 15:18:48.467380 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7gm8t" Feb 02 15:18:48 crc kubenswrapper[4733]: I0202 15:18:48.642301 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jk5b2" Feb 02 15:18:48 crc kubenswrapper[4733]: I0202 15:18:48.642354 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jk5b2" Feb 02 15:18:48 crc kubenswrapper[4733]: I0202 15:18:48.686753 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jk5b2" Feb 02 15:18:48 crc kubenswrapper[4733]: I0202 15:18:48.718443 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7gm8t" Feb 02 15:18:48 crc kubenswrapper[4733]: I0202 15:18:48.736751 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jk5b2" Feb 02 15:18:48 crc kubenswrapper[4733]: I0202 15:18:48.854765 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-m6bvv" Feb 02 15:18:48 crc kubenswrapper[4733]: I0202 15:18:48.854819 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-m6bvv" Feb 02 15:18:48 crc kubenswrapper[4733]: I0202 15:18:48.894572 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-m6bvv" Feb 02 15:18:49 crc kubenswrapper[4733]: I0202 15:18:49.737971 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-m6bvv" Feb 02 15:18:50 crc kubenswrapper[4733]: I0202 15:18:50.418739 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-v5t9s" Feb 02 15:18:50 crc kubenswrapper[4733]: I0202 15:18:50.419099 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-v5t9s" Feb 02 15:18:50 crc kubenswrapper[4733]: I0202 15:18:50.480673 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-v5t9s" Feb 02 15:18:50 crc kubenswrapper[4733]: I0202 15:18:50.725143 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-v5t9s" Feb 02 15:18:50 crc kubenswrapper[4733]: I0202 15:18:50.876632 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-76wkk" Feb 02 15:18:50 crc kubenswrapper[4733]: I0202 15:18:50.943070 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-76wkk" Feb 02 15:18:51 crc kubenswrapper[4733]: I0202 15:18:51.089721 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-m6bvv"] Feb 02 15:18:51 crc kubenswrapper[4733]: I0202 15:18:51.689453 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-m6bvv" podUID="5f14a53f-b3e0-46ab-af53-2dfb4041498d" containerName="registry-server" containerID="cri-o://c31ab2851ff2d0321b9b9a7e33773d9cd0cb7f4a078829841f34c98e26972eba" gracePeriod=2 Feb 02 15:18:51 crc kubenswrapper[4733]: I0202 15:18:51.700567 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9fckn" Feb 02 15:18:51 crc kubenswrapper[4733]: I0202 15:18:51.743203 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9fckn" Feb 02 15:18:52 crc kubenswrapper[4733]: I0202 15:18:52.075284 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-frkzg" Feb 02 15:18:52 crc kubenswrapper[4733]: I0202 15:18:52.128448 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-frkzg" Feb 02 15:18:52 crc kubenswrapper[4733]: I0202 15:18:52.620411 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m6bvv" Feb 02 15:18:52 crc kubenswrapper[4733]: I0202 15:18:52.695498 4733 generic.go:334] "Generic (PLEG): container finished" podID="5f14a53f-b3e0-46ab-af53-2dfb4041498d" containerID="c31ab2851ff2d0321b9b9a7e33773d9cd0cb7f4a078829841f34c98e26972eba" exitCode=0 Feb 02 15:18:52 crc kubenswrapper[4733]: I0202 15:18:52.695592 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m6bvv" Feb 02 15:18:52 crc kubenswrapper[4733]: I0202 15:18:52.695650 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m6bvv" event={"ID":"5f14a53f-b3e0-46ab-af53-2dfb4041498d","Type":"ContainerDied","Data":"c31ab2851ff2d0321b9b9a7e33773d9cd0cb7f4a078829841f34c98e26972eba"} Feb 02 15:18:52 crc kubenswrapper[4733]: I0202 15:18:52.695681 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m6bvv" event={"ID":"5f14a53f-b3e0-46ab-af53-2dfb4041498d","Type":"ContainerDied","Data":"1d1977da45425a58e7b2201f060d13186326918641a6805bbd30bd6d33494ca0"} Feb 02 15:18:52 crc kubenswrapper[4733]: I0202 15:18:52.695705 4733 scope.go:117] "RemoveContainer" containerID="c31ab2851ff2d0321b9b9a7e33773d9cd0cb7f4a078829841f34c98e26972eba" Feb 02 15:18:52 crc kubenswrapper[4733]: I0202 15:18:52.718609 4733 scope.go:117] "RemoveContainer" containerID="86bea7dcb5d2d734c5c76e7cc515dd97fb0b66c77788d02adee70f2855dd3727" Feb 02 15:18:52 crc kubenswrapper[4733]: I0202 15:18:52.736344 4733 scope.go:117] "RemoveContainer" containerID="81289aa37886a3f159d89678d35ecbb761830f4804aeb83caf3430344d5ef808" Feb 02 15:18:52 crc kubenswrapper[4733]: I0202 15:18:52.752651 4733 scope.go:117] "RemoveContainer" containerID="c31ab2851ff2d0321b9b9a7e33773d9cd0cb7f4a078829841f34c98e26972eba" Feb 02 15:18:52 crc kubenswrapper[4733]: E0202 15:18:52.753150 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c31ab2851ff2d0321b9b9a7e33773d9cd0cb7f4a078829841f34c98e26972eba\": container with ID starting with c31ab2851ff2d0321b9b9a7e33773d9cd0cb7f4a078829841f34c98e26972eba not found: ID does not exist" containerID="c31ab2851ff2d0321b9b9a7e33773d9cd0cb7f4a078829841f34c98e26972eba" Feb 02 15:18:52 crc kubenswrapper[4733]: I0202 15:18:52.753227 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c31ab2851ff2d0321b9b9a7e33773d9cd0cb7f4a078829841f34c98e26972eba"} err="failed to get container status \"c31ab2851ff2d0321b9b9a7e33773d9cd0cb7f4a078829841f34c98e26972eba\": rpc error: code = NotFound desc = could not find container \"c31ab2851ff2d0321b9b9a7e33773d9cd0cb7f4a078829841f34c98e26972eba\": container with ID starting with c31ab2851ff2d0321b9b9a7e33773d9cd0cb7f4a078829841f34c98e26972eba not found: ID does not exist" Feb 02 15:18:52 crc kubenswrapper[4733]: I0202 15:18:52.753282 4733 scope.go:117] "RemoveContainer" containerID="86bea7dcb5d2d734c5c76e7cc515dd97fb0b66c77788d02adee70f2855dd3727" Feb 02 15:18:52 crc kubenswrapper[4733]: E0202 15:18:52.753714 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86bea7dcb5d2d734c5c76e7cc515dd97fb0b66c77788d02adee70f2855dd3727\": container with ID starting with 86bea7dcb5d2d734c5c76e7cc515dd97fb0b66c77788d02adee70f2855dd3727 not found: ID does not exist" containerID="86bea7dcb5d2d734c5c76e7cc515dd97fb0b66c77788d02adee70f2855dd3727" Feb 02 15:18:52 crc kubenswrapper[4733]: I0202 15:18:52.753852 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86bea7dcb5d2d734c5c76e7cc515dd97fb0b66c77788d02adee70f2855dd3727"} err="failed to get container status \"86bea7dcb5d2d734c5c76e7cc515dd97fb0b66c77788d02adee70f2855dd3727\": rpc error: code = NotFound desc = could not find container \"86bea7dcb5d2d734c5c76e7cc515dd97fb0b66c77788d02adee70f2855dd3727\": container with ID starting with 86bea7dcb5d2d734c5c76e7cc515dd97fb0b66c77788d02adee70f2855dd3727 not found: ID does not exist" Feb 02 15:18:52 crc kubenswrapper[4733]: I0202 15:18:52.753972 4733 scope.go:117] "RemoveContainer" containerID="81289aa37886a3f159d89678d35ecbb761830f4804aeb83caf3430344d5ef808" Feb 02 15:18:52 crc kubenswrapper[4733]: E0202 15:18:52.754583 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81289aa37886a3f159d89678d35ecbb761830f4804aeb83caf3430344d5ef808\": container with ID starting with 81289aa37886a3f159d89678d35ecbb761830f4804aeb83caf3430344d5ef808 not found: ID does not exist" containerID="81289aa37886a3f159d89678d35ecbb761830f4804aeb83caf3430344d5ef808" Feb 02 15:18:52 crc kubenswrapper[4733]: I0202 15:18:52.754746 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81289aa37886a3f159d89678d35ecbb761830f4804aeb83caf3430344d5ef808"} err="failed to get container status \"81289aa37886a3f159d89678d35ecbb761830f4804aeb83caf3430344d5ef808\": rpc error: code = NotFound desc = could not find container \"81289aa37886a3f159d89678d35ecbb761830f4804aeb83caf3430344d5ef808\": container with ID starting with 81289aa37886a3f159d89678d35ecbb761830f4804aeb83caf3430344d5ef808 not found: ID does not exist" Feb 02 15:18:52 crc kubenswrapper[4733]: I0202 15:18:52.778305 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f14a53f-b3e0-46ab-af53-2dfb4041498d-utilities\") pod \"5f14a53f-b3e0-46ab-af53-2dfb4041498d\" (UID: \"5f14a53f-b3e0-46ab-af53-2dfb4041498d\") " Feb 02 15:18:52 crc kubenswrapper[4733]: I0202 15:18:52.778535 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d79mk\" (UniqueName: \"kubernetes.io/projected/5f14a53f-b3e0-46ab-af53-2dfb4041498d-kube-api-access-d79mk\") pod \"5f14a53f-b3e0-46ab-af53-2dfb4041498d\" (UID: \"5f14a53f-b3e0-46ab-af53-2dfb4041498d\") " Feb 02 15:18:52 crc kubenswrapper[4733]: I0202 15:18:52.778650 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f14a53f-b3e0-46ab-af53-2dfb4041498d-catalog-content\") pod \"5f14a53f-b3e0-46ab-af53-2dfb4041498d\" (UID: \"5f14a53f-b3e0-46ab-af53-2dfb4041498d\") " Feb 02 15:18:52 crc kubenswrapper[4733]: I0202 15:18:52.779269 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f14a53f-b3e0-46ab-af53-2dfb4041498d-utilities" (OuterVolumeSpecName: "utilities") pod "5f14a53f-b3e0-46ab-af53-2dfb4041498d" (UID: "5f14a53f-b3e0-46ab-af53-2dfb4041498d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:18:52 crc kubenswrapper[4733]: I0202 15:18:52.783766 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f14a53f-b3e0-46ab-af53-2dfb4041498d-kube-api-access-d79mk" (OuterVolumeSpecName: "kube-api-access-d79mk") pod "5f14a53f-b3e0-46ab-af53-2dfb4041498d" (UID: "5f14a53f-b3e0-46ab-af53-2dfb4041498d"). InnerVolumeSpecName "kube-api-access-d79mk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:18:52 crc kubenswrapper[4733]: I0202 15:18:52.823849 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f14a53f-b3e0-46ab-af53-2dfb4041498d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5f14a53f-b3e0-46ab-af53-2dfb4041498d" (UID: "5f14a53f-b3e0-46ab-af53-2dfb4041498d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:18:52 crc kubenswrapper[4733]: I0202 15:18:52.880072 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f14a53f-b3e0-46ab-af53-2dfb4041498d-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 15:18:52 crc kubenswrapper[4733]: I0202 15:18:52.880107 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d79mk\" (UniqueName: \"kubernetes.io/projected/5f14a53f-b3e0-46ab-af53-2dfb4041498d-kube-api-access-d79mk\") on node \"crc\" DevicePath \"\"" Feb 02 15:18:52 crc kubenswrapper[4733]: I0202 15:18:52.880118 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f14a53f-b3e0-46ab-af53-2dfb4041498d-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 15:18:53 crc kubenswrapper[4733]: I0202 15:18:53.019818 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-m6bvv"] Feb 02 15:18:53 crc kubenswrapper[4733]: I0202 15:18:53.023875 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-m6bvv"] Feb 02 15:18:53 crc kubenswrapper[4733]: I0202 15:18:53.263400 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f14a53f-b3e0-46ab-af53-2dfb4041498d" path="/var/lib/kubelet/pods/5f14a53f-b3e0-46ab-af53-2dfb4041498d/volumes" Feb 02 15:18:53 crc kubenswrapper[4733]: I0202 15:18:53.493674 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-76wkk"] Feb 02 15:18:53 crc kubenswrapper[4733]: I0202 15:18:53.493946 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-76wkk" podUID="316793cf-5542-4af4-b6ca-ad207ecc5009" containerName="registry-server" containerID="cri-o://996fc24a0c61b46b3872ab67c673a08e2060740f6c805e2dd382dbbcf213ebf6" gracePeriod=2 Feb 02 15:18:53 crc kubenswrapper[4733]: I0202 15:18:53.710498 4733 generic.go:334] "Generic (PLEG): container finished" podID="316793cf-5542-4af4-b6ca-ad207ecc5009" containerID="996fc24a0c61b46b3872ab67c673a08e2060740f6c805e2dd382dbbcf213ebf6" exitCode=0 Feb 02 15:18:53 crc kubenswrapper[4733]: I0202 15:18:53.710535 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-76wkk" event={"ID":"316793cf-5542-4af4-b6ca-ad207ecc5009","Type":"ContainerDied","Data":"996fc24a0c61b46b3872ab67c673a08e2060740f6c805e2dd382dbbcf213ebf6"} Feb 02 15:18:53 crc kubenswrapper[4733]: I0202 15:18:53.849078 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-76wkk" Feb 02 15:18:53 crc kubenswrapper[4733]: I0202 15:18:53.995911 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/316793cf-5542-4af4-b6ca-ad207ecc5009-utilities\") pod \"316793cf-5542-4af4-b6ca-ad207ecc5009\" (UID: \"316793cf-5542-4af4-b6ca-ad207ecc5009\") " Feb 02 15:18:53 crc kubenswrapper[4733]: I0202 15:18:53.995985 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/316793cf-5542-4af4-b6ca-ad207ecc5009-catalog-content\") pod \"316793cf-5542-4af4-b6ca-ad207ecc5009\" (UID: \"316793cf-5542-4af4-b6ca-ad207ecc5009\") " Feb 02 15:18:53 crc kubenswrapper[4733]: I0202 15:18:53.996011 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4jdp6\" (UniqueName: \"kubernetes.io/projected/316793cf-5542-4af4-b6ca-ad207ecc5009-kube-api-access-4jdp6\") pod \"316793cf-5542-4af4-b6ca-ad207ecc5009\" (UID: \"316793cf-5542-4af4-b6ca-ad207ecc5009\") " Feb 02 15:18:53 crc kubenswrapper[4733]: I0202 15:18:53.997542 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/316793cf-5542-4af4-b6ca-ad207ecc5009-utilities" (OuterVolumeSpecName: "utilities") pod "316793cf-5542-4af4-b6ca-ad207ecc5009" (UID: "316793cf-5542-4af4-b6ca-ad207ecc5009"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:18:54 crc kubenswrapper[4733]: I0202 15:18:54.000511 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/316793cf-5542-4af4-b6ca-ad207ecc5009-kube-api-access-4jdp6" (OuterVolumeSpecName: "kube-api-access-4jdp6") pod "316793cf-5542-4af4-b6ca-ad207ecc5009" (UID: "316793cf-5542-4af4-b6ca-ad207ecc5009"). InnerVolumeSpecName "kube-api-access-4jdp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:18:54 crc kubenswrapper[4733]: I0202 15:18:54.023313 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/316793cf-5542-4af4-b6ca-ad207ecc5009-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "316793cf-5542-4af4-b6ca-ad207ecc5009" (UID: "316793cf-5542-4af4-b6ca-ad207ecc5009"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:18:54 crc kubenswrapper[4733]: I0202 15:18:54.097070 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/316793cf-5542-4af4-b6ca-ad207ecc5009-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 15:18:54 crc kubenswrapper[4733]: I0202 15:18:54.097102 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/316793cf-5542-4af4-b6ca-ad207ecc5009-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 15:18:54 crc kubenswrapper[4733]: I0202 15:18:54.097117 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4jdp6\" (UniqueName: \"kubernetes.io/projected/316793cf-5542-4af4-b6ca-ad207ecc5009-kube-api-access-4jdp6\") on node \"crc\" DevicePath \"\"" Feb 02 15:18:54 crc kubenswrapper[4733]: I0202 15:18:54.717016 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-76wkk" event={"ID":"316793cf-5542-4af4-b6ca-ad207ecc5009","Type":"ContainerDied","Data":"a821ee12ee6448197d2deaf68c45ff5902304e01e0569e4a56d06a469dd005d6"} Feb 02 15:18:54 crc kubenswrapper[4733]: I0202 15:18:54.717079 4733 scope.go:117] "RemoveContainer" containerID="996fc24a0c61b46b3872ab67c673a08e2060740f6c805e2dd382dbbcf213ebf6" Feb 02 15:18:54 crc kubenswrapper[4733]: I0202 15:18:54.717100 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-76wkk" Feb 02 15:18:54 crc kubenswrapper[4733]: I0202 15:18:54.733065 4733 scope.go:117] "RemoveContainer" containerID="ce5e08f104b288ab3d5d9415126837e1d6fbf5e15520f2881b6c312c3bc56e3f" Feb 02 15:18:54 crc kubenswrapper[4733]: I0202 15:18:54.758253 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-76wkk"] Feb 02 15:18:54 crc kubenswrapper[4733]: I0202 15:18:54.758287 4733 scope.go:117] "RemoveContainer" containerID="d81c7c0de1b4b673862e5550c36f78c03a1bf16814b36eee87b777da041c41a9" Feb 02 15:18:54 crc kubenswrapper[4733]: I0202 15:18:54.763565 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-76wkk"] Feb 02 15:18:55 crc kubenswrapper[4733]: I0202 15:18:55.260438 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="316793cf-5542-4af4-b6ca-ad207ecc5009" path="/var/lib/kubelet/pods/316793cf-5542-4af4-b6ca-ad207ecc5009/volumes" Feb 02 15:18:55 crc kubenswrapper[4733]: I0202 15:18:55.894634 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-frkzg"] Feb 02 15:18:55 crc kubenswrapper[4733]: I0202 15:18:55.894849 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-frkzg" podUID="13882fb3-536e-4d85-ba06-4b7610a8d254" containerName="registry-server" containerID="cri-o://2862f3e13edc819755f766140152b4d94e06fb9ed85371e2f305d867a0443d57" gracePeriod=2 Feb 02 15:18:56 crc kubenswrapper[4733]: I0202 15:18:56.234527 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-frkzg" Feb 02 15:18:56 crc kubenswrapper[4733]: I0202 15:18:56.328698 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13882fb3-536e-4d85-ba06-4b7610a8d254-catalog-content\") pod \"13882fb3-536e-4d85-ba06-4b7610a8d254\" (UID: \"13882fb3-536e-4d85-ba06-4b7610a8d254\") " Feb 02 15:18:56 crc kubenswrapper[4733]: I0202 15:18:56.328759 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nwbjv\" (UniqueName: \"kubernetes.io/projected/13882fb3-536e-4d85-ba06-4b7610a8d254-kube-api-access-nwbjv\") pod \"13882fb3-536e-4d85-ba06-4b7610a8d254\" (UID: \"13882fb3-536e-4d85-ba06-4b7610a8d254\") " Feb 02 15:18:56 crc kubenswrapper[4733]: I0202 15:18:56.328835 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13882fb3-536e-4d85-ba06-4b7610a8d254-utilities\") pod \"13882fb3-536e-4d85-ba06-4b7610a8d254\" (UID: \"13882fb3-536e-4d85-ba06-4b7610a8d254\") " Feb 02 15:18:56 crc kubenswrapper[4733]: I0202 15:18:56.329662 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/13882fb3-536e-4d85-ba06-4b7610a8d254-utilities" (OuterVolumeSpecName: "utilities") pod "13882fb3-536e-4d85-ba06-4b7610a8d254" (UID: "13882fb3-536e-4d85-ba06-4b7610a8d254"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:18:56 crc kubenswrapper[4733]: I0202 15:18:56.333367 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13882fb3-536e-4d85-ba06-4b7610a8d254-kube-api-access-nwbjv" (OuterVolumeSpecName: "kube-api-access-nwbjv") pod "13882fb3-536e-4d85-ba06-4b7610a8d254" (UID: "13882fb3-536e-4d85-ba06-4b7610a8d254"). InnerVolumeSpecName "kube-api-access-nwbjv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:18:56 crc kubenswrapper[4733]: I0202 15:18:56.430469 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13882fb3-536e-4d85-ba06-4b7610a8d254-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 15:18:56 crc kubenswrapper[4733]: I0202 15:18:56.430510 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nwbjv\" (UniqueName: \"kubernetes.io/projected/13882fb3-536e-4d85-ba06-4b7610a8d254-kube-api-access-nwbjv\") on node \"crc\" DevicePath \"\"" Feb 02 15:18:56 crc kubenswrapper[4733]: I0202 15:18:56.445786 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/13882fb3-536e-4d85-ba06-4b7610a8d254-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "13882fb3-536e-4d85-ba06-4b7610a8d254" (UID: "13882fb3-536e-4d85-ba06-4b7610a8d254"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:18:56 crc kubenswrapper[4733]: I0202 15:18:56.531279 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13882fb3-536e-4d85-ba06-4b7610a8d254-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 15:18:56 crc kubenswrapper[4733]: I0202 15:18:56.732864 4733 generic.go:334] "Generic (PLEG): container finished" podID="13882fb3-536e-4d85-ba06-4b7610a8d254" containerID="2862f3e13edc819755f766140152b4d94e06fb9ed85371e2f305d867a0443d57" exitCode=0 Feb 02 15:18:56 crc kubenswrapper[4733]: I0202 15:18:56.732931 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frkzg" event={"ID":"13882fb3-536e-4d85-ba06-4b7610a8d254","Type":"ContainerDied","Data":"2862f3e13edc819755f766140152b4d94e06fb9ed85371e2f305d867a0443d57"} Feb 02 15:18:56 crc kubenswrapper[4733]: I0202 15:18:56.732983 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frkzg" event={"ID":"13882fb3-536e-4d85-ba06-4b7610a8d254","Type":"ContainerDied","Data":"aa6f2970f791eff9ab7f627c41cebdb0149aa9cc926f45e3bcf71bb62e0bc9f0"} Feb 02 15:18:56 crc kubenswrapper[4733]: I0202 15:18:56.733024 4733 scope.go:117] "RemoveContainer" containerID="2862f3e13edc819755f766140152b4d94e06fb9ed85371e2f305d867a0443d57" Feb 02 15:18:56 crc kubenswrapper[4733]: I0202 15:18:56.733041 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-frkzg" Feb 02 15:18:56 crc kubenswrapper[4733]: I0202 15:18:56.756058 4733 scope.go:117] "RemoveContainer" containerID="0afb3f4a8eaed5fa5ab6de21c6d633669c87ece98941e3071f58f97663f98f85" Feb 02 15:18:56 crc kubenswrapper[4733]: I0202 15:18:56.775751 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-frkzg"] Feb 02 15:18:56 crc kubenswrapper[4733]: I0202 15:18:56.781876 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-frkzg"] Feb 02 15:18:56 crc kubenswrapper[4733]: I0202 15:18:56.799755 4733 scope.go:117] "RemoveContainer" containerID="e055e0e861fc98fa5a4e9c5017041d1c15e921dff15f186e2cef0874301d7ce1" Feb 02 15:18:56 crc kubenswrapper[4733]: I0202 15:18:56.819653 4733 scope.go:117] "RemoveContainer" containerID="2862f3e13edc819755f766140152b4d94e06fb9ed85371e2f305d867a0443d57" Feb 02 15:18:56 crc kubenswrapper[4733]: E0202 15:18:56.820027 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2862f3e13edc819755f766140152b4d94e06fb9ed85371e2f305d867a0443d57\": container with ID starting with 2862f3e13edc819755f766140152b4d94e06fb9ed85371e2f305d867a0443d57 not found: ID does not exist" containerID="2862f3e13edc819755f766140152b4d94e06fb9ed85371e2f305d867a0443d57" Feb 02 15:18:56 crc kubenswrapper[4733]: I0202 15:18:56.820063 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2862f3e13edc819755f766140152b4d94e06fb9ed85371e2f305d867a0443d57"} err="failed to get container status \"2862f3e13edc819755f766140152b4d94e06fb9ed85371e2f305d867a0443d57\": rpc error: code = NotFound desc = could not find container \"2862f3e13edc819755f766140152b4d94e06fb9ed85371e2f305d867a0443d57\": container with ID starting with 2862f3e13edc819755f766140152b4d94e06fb9ed85371e2f305d867a0443d57 not found: ID does not exist" Feb 02 15:18:56 crc kubenswrapper[4733]: I0202 15:18:56.820100 4733 scope.go:117] "RemoveContainer" containerID="0afb3f4a8eaed5fa5ab6de21c6d633669c87ece98941e3071f58f97663f98f85" Feb 02 15:18:56 crc kubenswrapper[4733]: E0202 15:18:56.820394 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0afb3f4a8eaed5fa5ab6de21c6d633669c87ece98941e3071f58f97663f98f85\": container with ID starting with 0afb3f4a8eaed5fa5ab6de21c6d633669c87ece98941e3071f58f97663f98f85 not found: ID does not exist" containerID="0afb3f4a8eaed5fa5ab6de21c6d633669c87ece98941e3071f58f97663f98f85" Feb 02 15:18:56 crc kubenswrapper[4733]: I0202 15:18:56.820415 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0afb3f4a8eaed5fa5ab6de21c6d633669c87ece98941e3071f58f97663f98f85"} err="failed to get container status \"0afb3f4a8eaed5fa5ab6de21c6d633669c87ece98941e3071f58f97663f98f85\": rpc error: code = NotFound desc = could not find container \"0afb3f4a8eaed5fa5ab6de21c6d633669c87ece98941e3071f58f97663f98f85\": container with ID starting with 0afb3f4a8eaed5fa5ab6de21c6d633669c87ece98941e3071f58f97663f98f85 not found: ID does not exist" Feb 02 15:18:56 crc kubenswrapper[4733]: I0202 15:18:56.820430 4733 scope.go:117] "RemoveContainer" containerID="e055e0e861fc98fa5a4e9c5017041d1c15e921dff15f186e2cef0874301d7ce1" Feb 02 15:18:56 crc kubenswrapper[4733]: E0202 15:18:56.820766 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e055e0e861fc98fa5a4e9c5017041d1c15e921dff15f186e2cef0874301d7ce1\": container with ID starting with e055e0e861fc98fa5a4e9c5017041d1c15e921dff15f186e2cef0874301d7ce1 not found: ID does not exist" containerID="e055e0e861fc98fa5a4e9c5017041d1c15e921dff15f186e2cef0874301d7ce1" Feb 02 15:18:56 crc kubenswrapper[4733]: I0202 15:18:56.820800 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e055e0e861fc98fa5a4e9c5017041d1c15e921dff15f186e2cef0874301d7ce1"} err="failed to get container status \"e055e0e861fc98fa5a4e9c5017041d1c15e921dff15f186e2cef0874301d7ce1\": rpc error: code = NotFound desc = could not find container \"e055e0e861fc98fa5a4e9c5017041d1c15e921dff15f186e2cef0874301d7ce1\": container with ID starting with e055e0e861fc98fa5a4e9c5017041d1c15e921dff15f186e2cef0874301d7ce1 not found: ID does not exist" Feb 02 15:18:57 crc kubenswrapper[4733]: I0202 15:18:57.265409 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13882fb3-536e-4d85-ba06-4b7610a8d254" path="/var/lib/kubelet/pods/13882fb3-536e-4d85-ba06-4b7610a8d254/volumes" Feb 02 15:19:04 crc kubenswrapper[4733]: I0202 15:19:04.987789 4733 patch_prober.go:28] interesting pod/machine-config-daemon-zgsv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 15:19:04 crc kubenswrapper[4733]: I0202 15:19:04.988455 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 15:19:04 crc kubenswrapper[4733]: I0202 15:19:04.988519 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" Feb 02 15:19:04 crc kubenswrapper[4733]: I0202 15:19:04.989270 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d51318ad71b42f4cc8d376aa50e44580e68da89d83d80defb996329a2ee5c2ad"} pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 15:19:04 crc kubenswrapper[4733]: I0202 15:19:04.989359 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" containerName="machine-config-daemon" containerID="cri-o://d51318ad71b42f4cc8d376aa50e44580e68da89d83d80defb996329a2ee5c2ad" gracePeriod=600 Feb 02 15:19:05 crc kubenswrapper[4733]: I0202 15:19:05.793453 4733 generic.go:334] "Generic (PLEG): container finished" podID="83d0e810-ea5c-47aa-947b-7865b045bf94" containerID="d51318ad71b42f4cc8d376aa50e44580e68da89d83d80defb996329a2ee5c2ad" exitCode=0 Feb 02 15:19:05 crc kubenswrapper[4733]: I0202 15:19:05.793536 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" event={"ID":"83d0e810-ea5c-47aa-947b-7865b045bf94","Type":"ContainerDied","Data":"d51318ad71b42f4cc8d376aa50e44580e68da89d83d80defb996329a2ee5c2ad"} Feb 02 15:19:05 crc kubenswrapper[4733]: I0202 15:19:05.969823 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-4cq4j" podUID="7a9e1fe8-5666-4dae-baa9-df9fee6e21f1" containerName="oauth-openshift" containerID="cri-o://67433a8e25fd1c4786afeb674ca26aad6693a23348069352b1158a5edc1054a8" gracePeriod=15 Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.406807 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-4cq4j" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.444718 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-7f5b9fd94b-46nhc"] Feb 02 15:19:06 crc kubenswrapper[4733]: E0202 15:19:06.445015 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13882fb3-536e-4d85-ba06-4b7610a8d254" containerName="registry-server" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.445044 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="13882fb3-536e-4d85-ba06-4b7610a8d254" containerName="registry-server" Feb 02 15:19:06 crc kubenswrapper[4733]: E0202 15:19:06.445058 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13882fb3-536e-4d85-ba06-4b7610a8d254" containerName="extract-content" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.445067 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="13882fb3-536e-4d85-ba06-4b7610a8d254" containerName="extract-content" Feb 02 15:19:06 crc kubenswrapper[4733]: E0202 15:19:06.445081 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="316793cf-5542-4af4-b6ca-ad207ecc5009" containerName="registry-server" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.445090 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="316793cf-5542-4af4-b6ca-ad207ecc5009" containerName="registry-server" Feb 02 15:19:06 crc kubenswrapper[4733]: E0202 15:19:06.445103 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a9e1fe8-5666-4dae-baa9-df9fee6e21f1" containerName="oauth-openshift" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.445111 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a9e1fe8-5666-4dae-baa9-df9fee6e21f1" containerName="oauth-openshift" Feb 02 15:19:06 crc kubenswrapper[4733]: E0202 15:19:06.445126 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f14a53f-b3e0-46ab-af53-2dfb4041498d" containerName="extract-content" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.445134 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f14a53f-b3e0-46ab-af53-2dfb4041498d" containerName="extract-content" Feb 02 15:19:06 crc kubenswrapper[4733]: E0202 15:19:06.445148 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f14a53f-b3e0-46ab-af53-2dfb4041498d" containerName="registry-server" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.445179 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f14a53f-b3e0-46ab-af53-2dfb4041498d" containerName="registry-server" Feb 02 15:19:06 crc kubenswrapper[4733]: E0202 15:19:06.445197 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f14a53f-b3e0-46ab-af53-2dfb4041498d" containerName="extract-utilities" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.445206 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f14a53f-b3e0-46ab-af53-2dfb4041498d" containerName="extract-utilities" Feb 02 15:19:06 crc kubenswrapper[4733]: E0202 15:19:06.445216 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="316793cf-5542-4af4-b6ca-ad207ecc5009" containerName="extract-content" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.445225 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="316793cf-5542-4af4-b6ca-ad207ecc5009" containerName="extract-content" Feb 02 15:19:06 crc kubenswrapper[4733]: E0202 15:19:06.445240 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="316793cf-5542-4af4-b6ca-ad207ecc5009" containerName="extract-utilities" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.445249 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="316793cf-5542-4af4-b6ca-ad207ecc5009" containerName="extract-utilities" Feb 02 15:19:06 crc kubenswrapper[4733]: E0202 15:19:06.445260 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13882fb3-536e-4d85-ba06-4b7610a8d254" containerName="extract-utilities" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.445268 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="13882fb3-536e-4d85-ba06-4b7610a8d254" containerName="extract-utilities" Feb 02 15:19:06 crc kubenswrapper[4733]: E0202 15:19:06.445281 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d116a9fc-670f-4a77-b9ce-d6903d2c3d24" containerName="extract-utilities" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.445290 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="d116a9fc-670f-4a77-b9ce-d6903d2c3d24" containerName="extract-utilities" Feb 02 15:19:06 crc kubenswrapper[4733]: E0202 15:19:06.445299 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d116a9fc-670f-4a77-b9ce-d6903d2c3d24" containerName="extract-content" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.445307 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="d116a9fc-670f-4a77-b9ce-d6903d2c3d24" containerName="extract-content" Feb 02 15:19:06 crc kubenswrapper[4733]: E0202 15:19:06.445317 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d116a9fc-670f-4a77-b9ce-d6903d2c3d24" containerName="registry-server" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.445325 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="d116a9fc-670f-4a77-b9ce-d6903d2c3d24" containerName="registry-server" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.445443 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f14a53f-b3e0-46ab-af53-2dfb4041498d" containerName="registry-server" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.445458 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="316793cf-5542-4af4-b6ca-ad207ecc5009" containerName="registry-server" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.445474 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a9e1fe8-5666-4dae-baa9-df9fee6e21f1" containerName="oauth-openshift" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.445484 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="13882fb3-536e-4d85-ba06-4b7610a8d254" containerName="registry-server" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.445497 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="d116a9fc-670f-4a77-b9ce-d6903d2c3d24" containerName="registry-server" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.445894 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7f5b9fd94b-46nhc" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.464722 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7f5b9fd94b-46nhc"] Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.505125 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2ntr\" (UniqueName: \"kubernetes.io/projected/2753f32b-3e8f-475b-b898-0e3c145c33f4-kube-api-access-c2ntr\") pod \"oauth-openshift-7f5b9fd94b-46nhc\" (UID: \"2753f32b-3e8f-475b-b898-0e3c145c33f4\") " pod="openshift-authentication/oauth-openshift-7f5b9fd94b-46nhc" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.506672 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2753f32b-3e8f-475b-b898-0e3c145c33f4-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7f5b9fd94b-46nhc\" (UID: \"2753f32b-3e8f-475b-b898-0e3c145c33f4\") " pod="openshift-authentication/oauth-openshift-7f5b9fd94b-46nhc" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.506818 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2753f32b-3e8f-475b-b898-0e3c145c33f4-v4-0-config-system-service-ca\") pod \"oauth-openshift-7f5b9fd94b-46nhc\" (UID: \"2753f32b-3e8f-475b-b898-0e3c145c33f4\") " pod="openshift-authentication/oauth-openshift-7f5b9fd94b-46nhc" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.506982 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2753f32b-3e8f-475b-b898-0e3c145c33f4-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7f5b9fd94b-46nhc\" (UID: \"2753f32b-3e8f-475b-b898-0e3c145c33f4\") " pod="openshift-authentication/oauth-openshift-7f5b9fd94b-46nhc" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.507077 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2753f32b-3e8f-475b-b898-0e3c145c33f4-audit-dir\") pod \"oauth-openshift-7f5b9fd94b-46nhc\" (UID: \"2753f32b-3e8f-475b-b898-0e3c145c33f4\") " pod="openshift-authentication/oauth-openshift-7f5b9fd94b-46nhc" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.507199 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2753f32b-3e8f-475b-b898-0e3c145c33f4-v4-0-config-user-template-error\") pod \"oauth-openshift-7f5b9fd94b-46nhc\" (UID: \"2753f32b-3e8f-475b-b898-0e3c145c33f4\") " pod="openshift-authentication/oauth-openshift-7f5b9fd94b-46nhc" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.507386 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2753f32b-3e8f-475b-b898-0e3c145c33f4-v4-0-config-user-template-login\") pod \"oauth-openshift-7f5b9fd94b-46nhc\" (UID: \"2753f32b-3e8f-475b-b898-0e3c145c33f4\") " pod="openshift-authentication/oauth-openshift-7f5b9fd94b-46nhc" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.508964 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2753f32b-3e8f-475b-b898-0e3c145c33f4-v4-0-config-system-router-certs\") pod \"oauth-openshift-7f5b9fd94b-46nhc\" (UID: \"2753f32b-3e8f-475b-b898-0e3c145c33f4\") " pod="openshift-authentication/oauth-openshift-7f5b9fd94b-46nhc" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.509094 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2753f32b-3e8f-475b-b898-0e3c145c33f4-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7f5b9fd94b-46nhc\" (UID: \"2753f32b-3e8f-475b-b898-0e3c145c33f4\") " pod="openshift-authentication/oauth-openshift-7f5b9fd94b-46nhc" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.509238 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2753f32b-3e8f-475b-b898-0e3c145c33f4-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7f5b9fd94b-46nhc\" (UID: \"2753f32b-3e8f-475b-b898-0e3c145c33f4\") " pod="openshift-authentication/oauth-openshift-7f5b9fd94b-46nhc" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.509337 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2753f32b-3e8f-475b-b898-0e3c145c33f4-audit-policies\") pod \"oauth-openshift-7f5b9fd94b-46nhc\" (UID: \"2753f32b-3e8f-475b-b898-0e3c145c33f4\") " pod="openshift-authentication/oauth-openshift-7f5b9fd94b-46nhc" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.509497 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2753f32b-3e8f-475b-b898-0e3c145c33f4-v4-0-config-system-session\") pod \"oauth-openshift-7f5b9fd94b-46nhc\" (UID: \"2753f32b-3e8f-475b-b898-0e3c145c33f4\") " pod="openshift-authentication/oauth-openshift-7f5b9fd94b-46nhc" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.509550 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2753f32b-3e8f-475b-b898-0e3c145c33f4-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7f5b9fd94b-46nhc\" (UID: \"2753f32b-3e8f-475b-b898-0e3c145c33f4\") " pod="openshift-authentication/oauth-openshift-7f5b9fd94b-46nhc" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.509644 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2753f32b-3e8f-475b-b898-0e3c145c33f4-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7f5b9fd94b-46nhc\" (UID: \"2753f32b-3e8f-475b-b898-0e3c145c33f4\") " pod="openshift-authentication/oauth-openshift-7f5b9fd94b-46nhc" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.610987 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-user-template-login\") pod \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.611060 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-system-ocp-branding-template\") pod \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.611085 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-system-service-ca\") pod \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.611125 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-system-session\") pod \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.611152 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-user-template-provider-selection\") pod \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.611728 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pppg7\" (UniqueName: \"kubernetes.io/projected/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-kube-api-access-pppg7\") pod \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.611762 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-user-idp-0-file-data\") pod \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.611792 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-user-template-error\") pod \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.611815 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-system-serving-cert\") pod \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.611838 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-audit-dir\") pod \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.611860 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-system-trusted-ca-bundle\") pod \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.611895 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-audit-policies\") pod \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.611947 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-system-router-certs\") pod \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.611979 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-system-cliconfig\") pod \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\" (UID: \"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1\") " Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.612115 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2753f32b-3e8f-475b-b898-0e3c145c33f4-v4-0-config-system-router-certs\") pod \"oauth-openshift-7f5b9fd94b-46nhc\" (UID: \"2753f32b-3e8f-475b-b898-0e3c145c33f4\") " pod="openshift-authentication/oauth-openshift-7f5b9fd94b-46nhc" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.612145 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2753f32b-3e8f-475b-b898-0e3c145c33f4-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7f5b9fd94b-46nhc\" (UID: \"2753f32b-3e8f-475b-b898-0e3c145c33f4\") " pod="openshift-authentication/oauth-openshift-7f5b9fd94b-46nhc" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.612190 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2753f32b-3e8f-475b-b898-0e3c145c33f4-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7f5b9fd94b-46nhc\" (UID: \"2753f32b-3e8f-475b-b898-0e3c145c33f4\") " pod="openshift-authentication/oauth-openshift-7f5b9fd94b-46nhc" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.612214 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2753f32b-3e8f-475b-b898-0e3c145c33f4-audit-policies\") pod \"oauth-openshift-7f5b9fd94b-46nhc\" (UID: \"2753f32b-3e8f-475b-b898-0e3c145c33f4\") " pod="openshift-authentication/oauth-openshift-7f5b9fd94b-46nhc" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.612249 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2753f32b-3e8f-475b-b898-0e3c145c33f4-v4-0-config-system-session\") pod \"oauth-openshift-7f5b9fd94b-46nhc\" (UID: \"2753f32b-3e8f-475b-b898-0e3c145c33f4\") " pod="openshift-authentication/oauth-openshift-7f5b9fd94b-46nhc" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.612275 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2753f32b-3e8f-475b-b898-0e3c145c33f4-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7f5b9fd94b-46nhc\" (UID: \"2753f32b-3e8f-475b-b898-0e3c145c33f4\") " pod="openshift-authentication/oauth-openshift-7f5b9fd94b-46nhc" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.612298 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2753f32b-3e8f-475b-b898-0e3c145c33f4-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7f5b9fd94b-46nhc\" (UID: \"2753f32b-3e8f-475b-b898-0e3c145c33f4\") " pod="openshift-authentication/oauth-openshift-7f5b9fd94b-46nhc" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.612343 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2ntr\" (UniqueName: \"kubernetes.io/projected/2753f32b-3e8f-475b-b898-0e3c145c33f4-kube-api-access-c2ntr\") pod \"oauth-openshift-7f5b9fd94b-46nhc\" (UID: \"2753f32b-3e8f-475b-b898-0e3c145c33f4\") " pod="openshift-authentication/oauth-openshift-7f5b9fd94b-46nhc" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.612370 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2753f32b-3e8f-475b-b898-0e3c145c33f4-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7f5b9fd94b-46nhc\" (UID: \"2753f32b-3e8f-475b-b898-0e3c145c33f4\") " pod="openshift-authentication/oauth-openshift-7f5b9fd94b-46nhc" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.612396 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2753f32b-3e8f-475b-b898-0e3c145c33f4-v4-0-config-system-service-ca\") pod \"oauth-openshift-7f5b9fd94b-46nhc\" (UID: \"2753f32b-3e8f-475b-b898-0e3c145c33f4\") " pod="openshift-authentication/oauth-openshift-7f5b9fd94b-46nhc" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.612422 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2753f32b-3e8f-475b-b898-0e3c145c33f4-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7f5b9fd94b-46nhc\" (UID: \"2753f32b-3e8f-475b-b898-0e3c145c33f4\") " pod="openshift-authentication/oauth-openshift-7f5b9fd94b-46nhc" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.612447 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2753f32b-3e8f-475b-b898-0e3c145c33f4-audit-dir\") pod \"oauth-openshift-7f5b9fd94b-46nhc\" (UID: \"2753f32b-3e8f-475b-b898-0e3c145c33f4\") " pod="openshift-authentication/oauth-openshift-7f5b9fd94b-46nhc" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.612472 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2753f32b-3e8f-475b-b898-0e3c145c33f4-v4-0-config-user-template-error\") pod \"oauth-openshift-7f5b9fd94b-46nhc\" (UID: \"2753f32b-3e8f-475b-b898-0e3c145c33f4\") " pod="openshift-authentication/oauth-openshift-7f5b9fd94b-46nhc" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.612514 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2753f32b-3e8f-475b-b898-0e3c145c33f4-v4-0-config-user-template-login\") pod \"oauth-openshift-7f5b9fd94b-46nhc\" (UID: \"2753f32b-3e8f-475b-b898-0e3c145c33f4\") " pod="openshift-authentication/oauth-openshift-7f5b9fd94b-46nhc" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.613567 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "7a9e1fe8-5666-4dae-baa9-df9fee6e21f1" (UID: "7a9e1fe8-5666-4dae-baa9-df9fee6e21f1"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.613679 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "7a9e1fe8-5666-4dae-baa9-df9fee6e21f1" (UID: "7a9e1fe8-5666-4dae-baa9-df9fee6e21f1"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.613861 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2753f32b-3e8f-475b-b898-0e3c145c33f4-audit-policies\") pod \"oauth-openshift-7f5b9fd94b-46nhc\" (UID: \"2753f32b-3e8f-475b-b898-0e3c145c33f4\") " pod="openshift-authentication/oauth-openshift-7f5b9fd94b-46nhc" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.614343 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "7a9e1fe8-5666-4dae-baa9-df9fee6e21f1" (UID: "7a9e1fe8-5666-4dae-baa9-df9fee6e21f1"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.617839 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "7a9e1fe8-5666-4dae-baa9-df9fee6e21f1" (UID: "7a9e1fe8-5666-4dae-baa9-df9fee6e21f1"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.618534 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "7a9e1fe8-5666-4dae-baa9-df9fee6e21f1" (UID: "7a9e1fe8-5666-4dae-baa9-df9fee6e21f1"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.618824 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2753f32b-3e8f-475b-b898-0e3c145c33f4-v4-0-config-system-router-certs\") pod \"oauth-openshift-7f5b9fd94b-46nhc\" (UID: \"2753f32b-3e8f-475b-b898-0e3c145c33f4\") " pod="openshift-authentication/oauth-openshift-7f5b9fd94b-46nhc" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.619257 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-kube-api-access-pppg7" (OuterVolumeSpecName: "kube-api-access-pppg7") pod "7a9e1fe8-5666-4dae-baa9-df9fee6e21f1" (UID: "7a9e1fe8-5666-4dae-baa9-df9fee6e21f1"). InnerVolumeSpecName "kube-api-access-pppg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.619513 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "7a9e1fe8-5666-4dae-baa9-df9fee6e21f1" (UID: "7a9e1fe8-5666-4dae-baa9-df9fee6e21f1"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.619596 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2753f32b-3e8f-475b-b898-0e3c145c33f4-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7f5b9fd94b-46nhc\" (UID: \"2753f32b-3e8f-475b-b898-0e3c145c33f4\") " pod="openshift-authentication/oauth-openshift-7f5b9fd94b-46nhc" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.619925 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "7a9e1fe8-5666-4dae-baa9-df9fee6e21f1" (UID: "7a9e1fe8-5666-4dae-baa9-df9fee6e21f1"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.620122 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "7a9e1fe8-5666-4dae-baa9-df9fee6e21f1" (UID: "7a9e1fe8-5666-4dae-baa9-df9fee6e21f1"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.620322 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2753f32b-3e8f-475b-b898-0e3c145c33f4-v4-0-config-user-template-login\") pod \"oauth-openshift-7f5b9fd94b-46nhc\" (UID: \"2753f32b-3e8f-475b-b898-0e3c145c33f4\") " pod="openshift-authentication/oauth-openshift-7f5b9fd94b-46nhc" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.620456 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2753f32b-3e8f-475b-b898-0e3c145c33f4-audit-dir\") pod \"oauth-openshift-7f5b9fd94b-46nhc\" (UID: \"2753f32b-3e8f-475b-b898-0e3c145c33f4\") " pod="openshift-authentication/oauth-openshift-7f5b9fd94b-46nhc" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.620914 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "7a9e1fe8-5666-4dae-baa9-df9fee6e21f1" (UID: "7a9e1fe8-5666-4dae-baa9-df9fee6e21f1"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.621354 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2753f32b-3e8f-475b-b898-0e3c145c33f4-v4-0-config-system-service-ca\") pod \"oauth-openshift-7f5b9fd94b-46nhc\" (UID: \"2753f32b-3e8f-475b-b898-0e3c145c33f4\") " pod="openshift-authentication/oauth-openshift-7f5b9fd94b-46nhc" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.620942 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "7a9e1fe8-5666-4dae-baa9-df9fee6e21f1" (UID: "7a9e1fe8-5666-4dae-baa9-df9fee6e21f1"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.621869 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "7a9e1fe8-5666-4dae-baa9-df9fee6e21f1" (UID: "7a9e1fe8-5666-4dae-baa9-df9fee6e21f1"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.621372 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2753f32b-3e8f-475b-b898-0e3c145c33f4-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7f5b9fd94b-46nhc\" (UID: \"2753f32b-3e8f-475b-b898-0e3c145c33f4\") " pod="openshift-authentication/oauth-openshift-7f5b9fd94b-46nhc" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.622378 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "7a9e1fe8-5666-4dae-baa9-df9fee6e21f1" (UID: "7a9e1fe8-5666-4dae-baa9-df9fee6e21f1"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.623069 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2753f32b-3e8f-475b-b898-0e3c145c33f4-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7f5b9fd94b-46nhc\" (UID: \"2753f32b-3e8f-475b-b898-0e3c145c33f4\") " pod="openshift-authentication/oauth-openshift-7f5b9fd94b-46nhc" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.623091 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "7a9e1fe8-5666-4dae-baa9-df9fee6e21f1" (UID: "7a9e1fe8-5666-4dae-baa9-df9fee6e21f1"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.625790 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2753f32b-3e8f-475b-b898-0e3c145c33f4-v4-0-config-system-session\") pod \"oauth-openshift-7f5b9fd94b-46nhc\" (UID: \"2753f32b-3e8f-475b-b898-0e3c145c33f4\") " pod="openshift-authentication/oauth-openshift-7f5b9fd94b-46nhc" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.625817 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2753f32b-3e8f-475b-b898-0e3c145c33f4-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7f5b9fd94b-46nhc\" (UID: \"2753f32b-3e8f-475b-b898-0e3c145c33f4\") " pod="openshift-authentication/oauth-openshift-7f5b9fd94b-46nhc" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.625804 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2753f32b-3e8f-475b-b898-0e3c145c33f4-v4-0-config-user-template-error\") pod \"oauth-openshift-7f5b9fd94b-46nhc\" (UID: \"2753f32b-3e8f-475b-b898-0e3c145c33f4\") " pod="openshift-authentication/oauth-openshift-7f5b9fd94b-46nhc" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.626899 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2753f32b-3e8f-475b-b898-0e3c145c33f4-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7f5b9fd94b-46nhc\" (UID: \"2753f32b-3e8f-475b-b898-0e3c145c33f4\") " pod="openshift-authentication/oauth-openshift-7f5b9fd94b-46nhc" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.627760 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2753f32b-3e8f-475b-b898-0e3c145c33f4-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7f5b9fd94b-46nhc\" (UID: \"2753f32b-3e8f-475b-b898-0e3c145c33f4\") " pod="openshift-authentication/oauth-openshift-7f5b9fd94b-46nhc" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.640923 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2ntr\" (UniqueName: \"kubernetes.io/projected/2753f32b-3e8f-475b-b898-0e3c145c33f4-kube-api-access-c2ntr\") pod \"oauth-openshift-7f5b9fd94b-46nhc\" (UID: \"2753f32b-3e8f-475b-b898-0e3c145c33f4\") " pod="openshift-authentication/oauth-openshift-7f5b9fd94b-46nhc" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.713808 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.713847 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.713857 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.713867 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.713877 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pppg7\" (UniqueName: \"kubernetes.io/projected/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-kube-api-access-pppg7\") on node \"crc\" DevicePath \"\"" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.713886 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.713897 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.713907 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.713918 4733 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-audit-dir\") on node \"crc\" DevicePath \"\"" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.713927 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.713937 4733 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.713947 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.713957 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.713967 4733 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.795254 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7f5b9fd94b-46nhc" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.803285 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" event={"ID":"83d0e810-ea5c-47aa-947b-7865b045bf94","Type":"ContainerStarted","Data":"c5ef374155c9a13227e85afc96d240c8534badb0e00280dca27672b354c63947"} Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.806480 4733 generic.go:334] "Generic (PLEG): container finished" podID="7a9e1fe8-5666-4dae-baa9-df9fee6e21f1" containerID="67433a8e25fd1c4786afeb674ca26aad6693a23348069352b1158a5edc1054a8" exitCode=0 Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.806536 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-4cq4j" event={"ID":"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1","Type":"ContainerDied","Data":"67433a8e25fd1c4786afeb674ca26aad6693a23348069352b1158a5edc1054a8"} Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.806567 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-4cq4j" event={"ID":"7a9e1fe8-5666-4dae-baa9-df9fee6e21f1","Type":"ContainerDied","Data":"953ec4cf5eb753fc046706b992d748a9b4c63d57a2b621951ceff34fcbac7021"} Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.806624 4733 scope.go:117] "RemoveContainer" containerID="67433a8e25fd1c4786afeb674ca26aad6693a23348069352b1158a5edc1054a8" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.806774 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-4cq4j" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.861430 4733 scope.go:117] "RemoveContainer" containerID="67433a8e25fd1c4786afeb674ca26aad6693a23348069352b1158a5edc1054a8" Feb 02 15:19:06 crc kubenswrapper[4733]: E0202 15:19:06.862283 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67433a8e25fd1c4786afeb674ca26aad6693a23348069352b1158a5edc1054a8\": container with ID starting with 67433a8e25fd1c4786afeb674ca26aad6693a23348069352b1158a5edc1054a8 not found: ID does not exist" containerID="67433a8e25fd1c4786afeb674ca26aad6693a23348069352b1158a5edc1054a8" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.862320 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67433a8e25fd1c4786afeb674ca26aad6693a23348069352b1158a5edc1054a8"} err="failed to get container status \"67433a8e25fd1c4786afeb674ca26aad6693a23348069352b1158a5edc1054a8\": rpc error: code = NotFound desc = could not find container \"67433a8e25fd1c4786afeb674ca26aad6693a23348069352b1158a5edc1054a8\": container with ID starting with 67433a8e25fd1c4786afeb674ca26aad6693a23348069352b1158a5edc1054a8 not found: ID does not exist" Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.874989 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-4cq4j"] Feb 02 15:19:06 crc kubenswrapper[4733]: I0202 15:19:06.878889 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-4cq4j"] Feb 02 15:19:07 crc kubenswrapper[4733]: I0202 15:19:07.048709 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7f5b9fd94b-46nhc"] Feb 02 15:19:07 crc kubenswrapper[4733]: W0202 15:19:07.053267 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2753f32b_3e8f_475b_b898_0e3c145c33f4.slice/crio-438162a086ff59b924ae8b0bc2702202a31c7f8c2fe3bfecffaac490f6ea5002 WatchSource:0}: Error finding container 438162a086ff59b924ae8b0bc2702202a31c7f8c2fe3bfecffaac490f6ea5002: Status 404 returned error can't find the container with id 438162a086ff59b924ae8b0bc2702202a31c7f8c2fe3bfecffaac490f6ea5002 Feb 02 15:19:07 crc kubenswrapper[4733]: I0202 15:19:07.263475 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a9e1fe8-5666-4dae-baa9-df9fee6e21f1" path="/var/lib/kubelet/pods/7a9e1fe8-5666-4dae-baa9-df9fee6e21f1/volumes" Feb 02 15:19:07 crc kubenswrapper[4733]: I0202 15:19:07.816585 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7f5b9fd94b-46nhc" event={"ID":"2753f32b-3e8f-475b-b898-0e3c145c33f4","Type":"ContainerStarted","Data":"c0c7f13e6d5904262eb76498c4d2e7e59b90c98519e1156a9648358acdef5488"} Feb 02 15:19:07 crc kubenswrapper[4733]: I0202 15:19:07.816948 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7f5b9fd94b-46nhc" event={"ID":"2753f32b-3e8f-475b-b898-0e3c145c33f4","Type":"ContainerStarted","Data":"438162a086ff59b924ae8b0bc2702202a31c7f8c2fe3bfecffaac490f6ea5002"} Feb 02 15:19:07 crc kubenswrapper[4733]: I0202 15:19:07.850318 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-7f5b9fd94b-46nhc" podStartSLOduration=27.850287905 podStartE2EDuration="27.850287905s" podCreationTimestamp="2026-02-02 15:18:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:19:07.849007402 +0000 UTC m=+231.300468790" watchObservedRunningTime="2026-02-02 15:19:07.850287905 +0000 UTC m=+231.301749303" Feb 02 15:19:08 crc kubenswrapper[4733]: I0202 15:19:08.820861 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-7f5b9fd94b-46nhc" Feb 02 15:19:08 crc kubenswrapper[4733]: I0202 15:19:08.826704 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-7f5b9fd94b-46nhc" Feb 02 15:19:16 crc kubenswrapper[4733]: E0202 15:19:16.945343 4733 file.go:109] "Unable to process watch event" err="can't process config file \"/etc/kubernetes/manifests/kube-apiserver-startup-monitor-pod.yaml\": /etc/kubernetes/manifests/kube-apiserver-startup-monitor-pod.yaml: couldn't parse as pod(Object 'Kind' is missing in 'null'), please check config file" Feb 02 15:19:16 crc kubenswrapper[4733]: I0202 15:19:16.947791 4733 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 02 15:19:16 crc kubenswrapper[4733]: I0202 15:19:16.949148 4733 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 02 15:19:16 crc kubenswrapper[4733]: I0202 15:19:16.949420 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 15:19:16 crc kubenswrapper[4733]: I0202 15:19:16.949929 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://f037e379a109ce230098e3ca223fd1dea3149de3ebe601f56e2bb295f3093a3c" gracePeriod=15 Feb 02 15:19:16 crc kubenswrapper[4733]: I0202 15:19:16.949971 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://f6fdd19dac506f2d5d7eedc44ec2cda92a9381b73e91a45d949e4fcde4dfc283" gracePeriod=15 Feb 02 15:19:16 crc kubenswrapper[4733]: I0202 15:19:16.950023 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://226f5949820d4f498dbf6fa251c6edf1c6356900bd946e6afae2c29042b1c835" gracePeriod=15 Feb 02 15:19:16 crc kubenswrapper[4733]: I0202 15:19:16.949874 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://13965f5abcfb0a36c33310518955f192673e630a1aaba94b49400026de5bc2fc" gracePeriod=15 Feb 02 15:19:16 crc kubenswrapper[4733]: I0202 15:19:16.949813 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://fd383d69b9b27d52fa8672ff6e0294a74876d676973bdc0ca95125fef5f2d085" gracePeriod=15 Feb 02 15:19:16 crc kubenswrapper[4733]: I0202 15:19:16.950821 4733 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 02 15:19:16 crc kubenswrapper[4733]: E0202 15:19:16.951505 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 02 15:19:16 crc kubenswrapper[4733]: I0202 15:19:16.951547 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 02 15:19:16 crc kubenswrapper[4733]: E0202 15:19:16.951578 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Feb 02 15:19:16 crc kubenswrapper[4733]: I0202 15:19:16.951594 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Feb 02 15:19:16 crc kubenswrapper[4733]: E0202 15:19:16.951610 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 02 15:19:16 crc kubenswrapper[4733]: I0202 15:19:16.951623 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 02 15:19:16 crc kubenswrapper[4733]: E0202 15:19:16.951642 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 02 15:19:16 crc kubenswrapper[4733]: I0202 15:19:16.951654 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 02 15:19:16 crc kubenswrapper[4733]: E0202 15:19:16.951674 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 02 15:19:16 crc kubenswrapper[4733]: I0202 15:19:16.951686 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 02 15:19:16 crc kubenswrapper[4733]: E0202 15:19:16.951712 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 02 15:19:16 crc kubenswrapper[4733]: I0202 15:19:16.951724 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 02 15:19:16 crc kubenswrapper[4733]: E0202 15:19:16.951742 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 02 15:19:16 crc kubenswrapper[4733]: I0202 15:19:16.951756 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 02 15:19:16 crc kubenswrapper[4733]: I0202 15:19:16.951979 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 02 15:19:16 crc kubenswrapper[4733]: I0202 15:19:16.952011 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 02 15:19:16 crc kubenswrapper[4733]: I0202 15:19:16.952037 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 02 15:19:16 crc kubenswrapper[4733]: I0202 15:19:16.952063 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 02 15:19:16 crc kubenswrapper[4733]: I0202 15:19:16.952084 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 02 15:19:16 crc kubenswrapper[4733]: I0202 15:19:16.952504 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 02 15:19:16 crc kubenswrapper[4733]: I0202 15:19:16.961563 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 15:19:16 crc kubenswrapper[4733]: I0202 15:19:16.961913 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 15:19:16 crc kubenswrapper[4733]: I0202 15:19:16.962097 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 15:19:16 crc kubenswrapper[4733]: I0202 15:19:16.962283 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 15:19:16 crc kubenswrapper[4733]: I0202 15:19:16.962392 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 15:19:17 crc kubenswrapper[4733]: I0202 15:19:16.998714 4733 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Readiness probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Feb 02 15:19:17 crc kubenswrapper[4733]: [+]log ok Feb 02 15:19:17 crc kubenswrapper[4733]: [+]api-openshift-apiserver-available ok Feb 02 15:19:17 crc kubenswrapper[4733]: [+]api-openshift-oauth-apiserver-available ok Feb 02 15:19:17 crc kubenswrapper[4733]: [+]informer-sync ok Feb 02 15:19:17 crc kubenswrapper[4733]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Feb 02 15:19:17 crc kubenswrapper[4733]: [+]poststarthook/openshift.io-api-request-count-filter ok Feb 02 15:19:17 crc kubenswrapper[4733]: [+]poststarthook/openshift.io-startkubeinformers ok Feb 02 15:19:17 crc kubenswrapper[4733]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Feb 02 15:19:17 crc kubenswrapper[4733]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Feb 02 15:19:17 crc kubenswrapper[4733]: [+]poststarthook/start-apiserver-admission-initializer ok Feb 02 15:19:17 crc kubenswrapper[4733]: [+]poststarthook/generic-apiserver-start-informers ok Feb 02 15:19:17 crc kubenswrapper[4733]: [+]poststarthook/priority-and-fairness-config-consumer ok Feb 02 15:19:17 crc kubenswrapper[4733]: [+]poststarthook/priority-and-fairness-filter ok Feb 02 15:19:17 crc kubenswrapper[4733]: [+]poststarthook/storage-object-count-tracker-hook ok Feb 02 15:19:17 crc kubenswrapper[4733]: [+]poststarthook/start-apiextensions-informers ok Feb 02 15:19:17 crc kubenswrapper[4733]: [+]poststarthook/start-apiextensions-controllers ok Feb 02 15:19:17 crc kubenswrapper[4733]: [+]poststarthook/crd-informer-synced ok Feb 02 15:19:17 crc kubenswrapper[4733]: [+]poststarthook/start-system-namespaces-controller ok Feb 02 15:19:17 crc kubenswrapper[4733]: [+]poststarthook/start-cluster-authentication-info-controller ok Feb 02 15:19:17 crc kubenswrapper[4733]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Feb 02 15:19:17 crc kubenswrapper[4733]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Feb 02 15:19:17 crc kubenswrapper[4733]: [+]poststarthook/start-legacy-token-tracking-controller ok Feb 02 15:19:17 crc kubenswrapper[4733]: [+]poststarthook/start-service-ip-repair-controllers ok Feb 02 15:19:17 crc kubenswrapper[4733]: [+]poststarthook/rbac/bootstrap-roles ok Feb 02 15:19:17 crc kubenswrapper[4733]: [+]poststarthook/scheduling/bootstrap-system-priority-classes ok Feb 02 15:19:17 crc kubenswrapper[4733]: [+]poststarthook/priority-and-fairness-config-producer ok Feb 02 15:19:17 crc kubenswrapper[4733]: [+]poststarthook/bootstrap-controller ok Feb 02 15:19:17 crc kubenswrapper[4733]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Feb 02 15:19:17 crc kubenswrapper[4733]: [+]poststarthook/start-kube-aggregator-informers ok Feb 02 15:19:17 crc kubenswrapper[4733]: [+]poststarthook/apiservice-status-local-available-controller ok Feb 02 15:19:17 crc kubenswrapper[4733]: [+]poststarthook/apiservice-status-remote-available-controller ok Feb 02 15:19:17 crc kubenswrapper[4733]: [+]poststarthook/apiservice-registration-controller ok Feb 02 15:19:17 crc kubenswrapper[4733]: [+]poststarthook/apiservice-wait-for-first-sync ok Feb 02 15:19:17 crc kubenswrapper[4733]: [+]poststarthook/apiservice-discovery-controller ok Feb 02 15:19:17 crc kubenswrapper[4733]: [+]poststarthook/kube-apiserver-autoregistration ok Feb 02 15:19:17 crc kubenswrapper[4733]: [+]autoregister-completion ok Feb 02 15:19:17 crc kubenswrapper[4733]: [+]poststarthook/apiservice-openapi-controller ok Feb 02 15:19:17 crc kubenswrapper[4733]: [+]poststarthook/apiservice-openapiv3-controller ok Feb 02 15:19:17 crc kubenswrapper[4733]: [-]shutdown failed: reason withheld Feb 02 15:19:17 crc kubenswrapper[4733]: readyz check failed Feb 02 15:19:17 crc kubenswrapper[4733]: I0202 15:19:16.998795 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 02 15:19:17 crc kubenswrapper[4733]: E0202 15:19:16.999903 4733 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.2:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 15:19:17 crc kubenswrapper[4733]: I0202 15:19:17.063507 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 15:19:17 crc kubenswrapper[4733]: I0202 15:19:17.063604 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 15:19:17 crc kubenswrapper[4733]: I0202 15:19:17.063655 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 15:19:17 crc kubenswrapper[4733]: I0202 15:19:17.063677 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 15:19:17 crc kubenswrapper[4733]: I0202 15:19:17.063700 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 15:19:17 crc kubenswrapper[4733]: I0202 15:19:17.063766 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 15:19:17 crc kubenswrapper[4733]: I0202 15:19:17.063830 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 15:19:17 crc kubenswrapper[4733]: I0202 15:19:17.063934 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 15:19:17 crc kubenswrapper[4733]: I0202 15:19:17.064011 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 15:19:17 crc kubenswrapper[4733]: I0202 15:19:17.064100 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 15:19:17 crc kubenswrapper[4733]: I0202 15:19:17.064136 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 15:19:17 crc kubenswrapper[4733]: I0202 15:19:17.064146 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 15:19:17 crc kubenswrapper[4733]: I0202 15:19:17.064090 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 15:19:17 crc kubenswrapper[4733]: E0202 15:19:17.136210 4733 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.2:6443: connect: connection refused" Feb 02 15:19:17 crc kubenswrapper[4733]: E0202 15:19:17.136836 4733 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.2:6443: connect: connection refused" Feb 02 15:19:17 crc kubenswrapper[4733]: E0202 15:19:17.137029 4733 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.2:6443: connect: connection refused" Feb 02 15:19:17 crc kubenswrapper[4733]: E0202 15:19:17.137225 4733 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.2:6443: connect: connection refused" Feb 02 15:19:17 crc kubenswrapper[4733]: E0202 15:19:17.137405 4733 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.2:6443: connect: connection refused" Feb 02 15:19:17 crc kubenswrapper[4733]: I0202 15:19:17.137425 4733 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Feb 02 15:19:17 crc kubenswrapper[4733]: E0202 15:19:17.137596 4733 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.2:6443: connect: connection refused" interval="200ms" Feb 02 15:19:17 crc kubenswrapper[4733]: I0202 15:19:17.165378 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 15:19:17 crc kubenswrapper[4733]: I0202 15:19:17.165428 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 15:19:17 crc kubenswrapper[4733]: I0202 15:19:17.165464 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 15:19:17 crc kubenswrapper[4733]: I0202 15:19:17.165532 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 15:19:17 crc kubenswrapper[4733]: I0202 15:19:17.165566 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 15:19:17 crc kubenswrapper[4733]: I0202 15:19:17.165585 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 15:19:17 crc kubenswrapper[4733]: I0202 15:19:17.261209 4733 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.2:6443: connect: connection refused" Feb 02 15:19:17 crc kubenswrapper[4733]: I0202 15:19:17.300936 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 15:19:17 crc kubenswrapper[4733]: W0202 15:19:17.334865 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-89c5e3d2af035476b26c0c9e1a075ccf80f48b54592e06b1811300c1bc057182 WatchSource:0}: Error finding container 89c5e3d2af035476b26c0c9e1a075ccf80f48b54592e06b1811300c1bc057182: Status 404 returned error can't find the container with id 89c5e3d2af035476b26c0c9e1a075ccf80f48b54592e06b1811300c1bc057182 Feb 02 15:19:17 crc kubenswrapper[4733]: E0202 15:19:17.338461 4733 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.2:6443: connect: connection refused" interval="400ms" Feb 02 15:19:17 crc kubenswrapper[4733]: E0202 15:19:17.338789 4733 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.2:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.1890770e69673103 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-02 15:19:17.338222851 +0000 UTC m=+240.789684219,LastTimestamp:2026-02-02 15:19:17.338222851 +0000 UTC m=+240.789684219,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 02 15:19:17 crc kubenswrapper[4733]: E0202 15:19:17.739518 4733 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.2:6443: connect: connection refused" interval="800ms" Feb 02 15:19:17 crc kubenswrapper[4733]: I0202 15:19:17.875800 4733 generic.go:334] "Generic (PLEG): container finished" podID="272ea61e-a61b-40c8-8e4a-cdc9f3c95847" containerID="a501edad55227183ad5dfc3a0ac0dfec399c4b05bc14706420020507a35c6599" exitCode=0 Feb 02 15:19:17 crc kubenswrapper[4733]: I0202 15:19:17.875886 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"272ea61e-a61b-40c8-8e4a-cdc9f3c95847","Type":"ContainerDied","Data":"a501edad55227183ad5dfc3a0ac0dfec399c4b05bc14706420020507a35c6599"} Feb 02 15:19:17 crc kubenswrapper[4733]: I0202 15:19:17.876659 4733 status_manager.go:851] "Failed to get status for pod" podUID="272ea61e-a61b-40c8-8e4a-cdc9f3c95847" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.2:6443: connect: connection refused" Feb 02 15:19:17 crc kubenswrapper[4733]: I0202 15:19:17.879666 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 02 15:19:17 crc kubenswrapper[4733]: I0202 15:19:17.881437 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 02 15:19:17 crc kubenswrapper[4733]: I0202 15:19:17.882579 4733 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="13965f5abcfb0a36c33310518955f192673e630a1aaba94b49400026de5bc2fc" exitCode=0 Feb 02 15:19:17 crc kubenswrapper[4733]: I0202 15:19:17.882602 4733 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="226f5949820d4f498dbf6fa251c6edf1c6356900bd946e6afae2c29042b1c835" exitCode=0 Feb 02 15:19:17 crc kubenswrapper[4733]: I0202 15:19:17.882613 4733 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="f037e379a109ce230098e3ca223fd1dea3149de3ebe601f56e2bb295f3093a3c" exitCode=0 Feb 02 15:19:17 crc kubenswrapper[4733]: I0202 15:19:17.882622 4733 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="f6fdd19dac506f2d5d7eedc44ec2cda92a9381b73e91a45d949e4fcde4dfc283" exitCode=2 Feb 02 15:19:17 crc kubenswrapper[4733]: I0202 15:19:17.882725 4733 scope.go:117] "RemoveContainer" containerID="992f4198192cbe4b1c89d575aa4a20c29c1115fd5eaa8986c6ed70c1d6b143c4" Feb 02 15:19:17 crc kubenswrapper[4733]: I0202 15:19:17.884832 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"c5c8bdc71a33879c2f3adefae66e3670d35ca4234942866cf4c153aee739ccfa"} Feb 02 15:19:17 crc kubenswrapper[4733]: I0202 15:19:17.884865 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"89c5e3d2af035476b26c0c9e1a075ccf80f48b54592e06b1811300c1bc057182"} Feb 02 15:19:17 crc kubenswrapper[4733]: E0202 15:19:17.885474 4733 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.2:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 15:19:17 crc kubenswrapper[4733]: I0202 15:19:17.885812 4733 status_manager.go:851] "Failed to get status for pod" podUID="272ea61e-a61b-40c8-8e4a-cdc9f3c95847" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.2:6443: connect: connection refused" Feb 02 15:19:18 crc kubenswrapper[4733]: E0202 15:19:18.540812 4733 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.2:6443: connect: connection refused" interval="1.6s" Feb 02 15:19:18 crc kubenswrapper[4733]: I0202 15:19:18.903860 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 02 15:19:19 crc kubenswrapper[4733]: I0202 15:19:19.342726 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 02 15:19:19 crc kubenswrapper[4733]: I0202 15:19:19.344338 4733 status_manager.go:851] "Failed to get status for pod" podUID="272ea61e-a61b-40c8-8e4a-cdc9f3c95847" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.2:6443: connect: connection refused" Feb 02 15:19:19 crc kubenswrapper[4733]: I0202 15:19:19.350710 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 02 15:19:19 crc kubenswrapper[4733]: I0202 15:19:19.351934 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 15:19:19 crc kubenswrapper[4733]: I0202 15:19:19.352770 4733 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.2:6443: connect: connection refused" Feb 02 15:19:19 crc kubenswrapper[4733]: I0202 15:19:19.353224 4733 status_manager.go:851] "Failed to get status for pod" podUID="272ea61e-a61b-40c8-8e4a-cdc9f3c95847" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.2:6443: connect: connection refused" Feb 02 15:19:19 crc kubenswrapper[4733]: I0202 15:19:19.507494 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/272ea61e-a61b-40c8-8e4a-cdc9f3c95847-var-lock\") pod \"272ea61e-a61b-40c8-8e4a-cdc9f3c95847\" (UID: \"272ea61e-a61b-40c8-8e4a-cdc9f3c95847\") " Feb 02 15:19:19 crc kubenswrapper[4733]: I0202 15:19:19.507560 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 02 15:19:19 crc kubenswrapper[4733]: I0202 15:19:19.507601 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/272ea61e-a61b-40c8-8e4a-cdc9f3c95847-kubelet-dir\") pod \"272ea61e-a61b-40c8-8e4a-cdc9f3c95847\" (UID: \"272ea61e-a61b-40c8-8e4a-cdc9f3c95847\") " Feb 02 15:19:19 crc kubenswrapper[4733]: I0202 15:19:19.507687 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 02 15:19:19 crc kubenswrapper[4733]: I0202 15:19:19.507737 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 02 15:19:19 crc kubenswrapper[4733]: I0202 15:19:19.507844 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/272ea61e-a61b-40c8-8e4a-cdc9f3c95847-kube-api-access\") pod \"272ea61e-a61b-40c8-8e4a-cdc9f3c95847\" (UID: \"272ea61e-a61b-40c8-8e4a-cdc9f3c95847\") " Feb 02 15:19:19 crc kubenswrapper[4733]: I0202 15:19:19.507860 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 15:19:19 crc kubenswrapper[4733]: I0202 15:19:19.507940 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/272ea61e-a61b-40c8-8e4a-cdc9f3c95847-var-lock" (OuterVolumeSpecName: "var-lock") pod "272ea61e-a61b-40c8-8e4a-cdc9f3c95847" (UID: "272ea61e-a61b-40c8-8e4a-cdc9f3c95847"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 15:19:19 crc kubenswrapper[4733]: I0202 15:19:19.507972 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 15:19:19 crc kubenswrapper[4733]: I0202 15:19:19.508004 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/272ea61e-a61b-40c8-8e4a-cdc9f3c95847-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "272ea61e-a61b-40c8-8e4a-cdc9f3c95847" (UID: "272ea61e-a61b-40c8-8e4a-cdc9f3c95847"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 15:19:19 crc kubenswrapper[4733]: I0202 15:19:19.508032 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 15:19:19 crc kubenswrapper[4733]: I0202 15:19:19.508239 4733 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Feb 02 15:19:19 crc kubenswrapper[4733]: I0202 15:19:19.508265 4733 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/272ea61e-a61b-40c8-8e4a-cdc9f3c95847-var-lock\") on node \"crc\" DevicePath \"\"" Feb 02 15:19:19 crc kubenswrapper[4733]: I0202 15:19:19.508281 4733 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 02 15:19:19 crc kubenswrapper[4733]: I0202 15:19:19.508298 4733 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/272ea61e-a61b-40c8-8e4a-cdc9f3c95847-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 02 15:19:19 crc kubenswrapper[4733]: I0202 15:19:19.508314 4733 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Feb 02 15:19:19 crc kubenswrapper[4733]: I0202 15:19:19.517679 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/272ea61e-a61b-40c8-8e4a-cdc9f3c95847-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "272ea61e-a61b-40c8-8e4a-cdc9f3c95847" (UID: "272ea61e-a61b-40c8-8e4a-cdc9f3c95847"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:19:19 crc kubenswrapper[4733]: I0202 15:19:19.609688 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/272ea61e-a61b-40c8-8e4a-cdc9f3c95847-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 02 15:19:19 crc kubenswrapper[4733]: E0202 15:19:19.787289 4733 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.2:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.1890770e69673103 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-02 15:19:17.338222851 +0000 UTC m=+240.789684219,LastTimestamp:2026-02-02 15:19:17.338222851 +0000 UTC m=+240.789684219,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 02 15:19:19 crc kubenswrapper[4733]: I0202 15:19:19.917101 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 02 15:19:19 crc kubenswrapper[4733]: I0202 15:19:19.917095 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"272ea61e-a61b-40c8-8e4a-cdc9f3c95847","Type":"ContainerDied","Data":"812ca4ef6743546483882c26c88426865bf39206c0ffc4e3b25472e6f5588151"} Feb 02 15:19:19 crc kubenswrapper[4733]: I0202 15:19:19.917260 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="812ca4ef6743546483882c26c88426865bf39206c0ffc4e3b25472e6f5588151" Feb 02 15:19:19 crc kubenswrapper[4733]: I0202 15:19:19.921857 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 02 15:19:19 crc kubenswrapper[4733]: I0202 15:19:19.923404 4733 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="fd383d69b9b27d52fa8672ff6e0294a74876d676973bdc0ca95125fef5f2d085" exitCode=0 Feb 02 15:19:19 crc kubenswrapper[4733]: I0202 15:19:19.923544 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 15:19:19 crc kubenswrapper[4733]: I0202 15:19:19.923565 4733 scope.go:117] "RemoveContainer" containerID="13965f5abcfb0a36c33310518955f192673e630a1aaba94b49400026de5bc2fc" Feb 02 15:19:19 crc kubenswrapper[4733]: I0202 15:19:19.950314 4733 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.2:6443: connect: connection refused" Feb 02 15:19:19 crc kubenswrapper[4733]: I0202 15:19:19.950995 4733 status_manager.go:851] "Failed to get status for pod" podUID="272ea61e-a61b-40c8-8e4a-cdc9f3c95847" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.2:6443: connect: connection refused" Feb 02 15:19:19 crc kubenswrapper[4733]: I0202 15:19:19.953411 4733 scope.go:117] "RemoveContainer" containerID="226f5949820d4f498dbf6fa251c6edf1c6356900bd946e6afae2c29042b1c835" Feb 02 15:19:19 crc kubenswrapper[4733]: I0202 15:19:19.955725 4733 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.2:6443: connect: connection refused" Feb 02 15:19:19 crc kubenswrapper[4733]: I0202 15:19:19.956206 4733 status_manager.go:851] "Failed to get status for pod" podUID="272ea61e-a61b-40c8-8e4a-cdc9f3c95847" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.2:6443: connect: connection refused" Feb 02 15:19:19 crc kubenswrapper[4733]: I0202 15:19:19.975376 4733 scope.go:117] "RemoveContainer" containerID="f037e379a109ce230098e3ca223fd1dea3149de3ebe601f56e2bb295f3093a3c" Feb 02 15:19:19 crc kubenswrapper[4733]: I0202 15:19:19.998450 4733 scope.go:117] "RemoveContainer" containerID="f6fdd19dac506f2d5d7eedc44ec2cda92a9381b73e91a45d949e4fcde4dfc283" Feb 02 15:19:20 crc kubenswrapper[4733]: I0202 15:19:20.024099 4733 scope.go:117] "RemoveContainer" containerID="fd383d69b9b27d52fa8672ff6e0294a74876d676973bdc0ca95125fef5f2d085" Feb 02 15:19:20 crc kubenswrapper[4733]: I0202 15:19:20.045528 4733 scope.go:117] "RemoveContainer" containerID="61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18" Feb 02 15:19:20 crc kubenswrapper[4733]: I0202 15:19:20.074216 4733 scope.go:117] "RemoveContainer" containerID="13965f5abcfb0a36c33310518955f192673e630a1aaba94b49400026de5bc2fc" Feb 02 15:19:20 crc kubenswrapper[4733]: E0202 15:19:20.075340 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13965f5abcfb0a36c33310518955f192673e630a1aaba94b49400026de5bc2fc\": container with ID starting with 13965f5abcfb0a36c33310518955f192673e630a1aaba94b49400026de5bc2fc not found: ID does not exist" containerID="13965f5abcfb0a36c33310518955f192673e630a1aaba94b49400026de5bc2fc" Feb 02 15:19:20 crc kubenswrapper[4733]: I0202 15:19:20.075398 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13965f5abcfb0a36c33310518955f192673e630a1aaba94b49400026de5bc2fc"} err="failed to get container status \"13965f5abcfb0a36c33310518955f192673e630a1aaba94b49400026de5bc2fc\": rpc error: code = NotFound desc = could not find container \"13965f5abcfb0a36c33310518955f192673e630a1aaba94b49400026de5bc2fc\": container with ID starting with 13965f5abcfb0a36c33310518955f192673e630a1aaba94b49400026de5bc2fc not found: ID does not exist" Feb 02 15:19:20 crc kubenswrapper[4733]: I0202 15:19:20.075480 4733 scope.go:117] "RemoveContainer" containerID="226f5949820d4f498dbf6fa251c6edf1c6356900bd946e6afae2c29042b1c835" Feb 02 15:19:20 crc kubenswrapper[4733]: E0202 15:19:20.075863 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"226f5949820d4f498dbf6fa251c6edf1c6356900bd946e6afae2c29042b1c835\": container with ID starting with 226f5949820d4f498dbf6fa251c6edf1c6356900bd946e6afae2c29042b1c835 not found: ID does not exist" containerID="226f5949820d4f498dbf6fa251c6edf1c6356900bd946e6afae2c29042b1c835" Feb 02 15:19:20 crc kubenswrapper[4733]: I0202 15:19:20.075917 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"226f5949820d4f498dbf6fa251c6edf1c6356900bd946e6afae2c29042b1c835"} err="failed to get container status \"226f5949820d4f498dbf6fa251c6edf1c6356900bd946e6afae2c29042b1c835\": rpc error: code = NotFound desc = could not find container \"226f5949820d4f498dbf6fa251c6edf1c6356900bd946e6afae2c29042b1c835\": container with ID starting with 226f5949820d4f498dbf6fa251c6edf1c6356900bd946e6afae2c29042b1c835 not found: ID does not exist" Feb 02 15:19:20 crc kubenswrapper[4733]: I0202 15:19:20.075953 4733 scope.go:117] "RemoveContainer" containerID="f037e379a109ce230098e3ca223fd1dea3149de3ebe601f56e2bb295f3093a3c" Feb 02 15:19:20 crc kubenswrapper[4733]: E0202 15:19:20.076884 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f037e379a109ce230098e3ca223fd1dea3149de3ebe601f56e2bb295f3093a3c\": container with ID starting with f037e379a109ce230098e3ca223fd1dea3149de3ebe601f56e2bb295f3093a3c not found: ID does not exist" containerID="f037e379a109ce230098e3ca223fd1dea3149de3ebe601f56e2bb295f3093a3c" Feb 02 15:19:20 crc kubenswrapper[4733]: I0202 15:19:20.076913 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f037e379a109ce230098e3ca223fd1dea3149de3ebe601f56e2bb295f3093a3c"} err="failed to get container status \"f037e379a109ce230098e3ca223fd1dea3149de3ebe601f56e2bb295f3093a3c\": rpc error: code = NotFound desc = could not find container \"f037e379a109ce230098e3ca223fd1dea3149de3ebe601f56e2bb295f3093a3c\": container with ID starting with f037e379a109ce230098e3ca223fd1dea3149de3ebe601f56e2bb295f3093a3c not found: ID does not exist" Feb 02 15:19:20 crc kubenswrapper[4733]: I0202 15:19:20.077004 4733 scope.go:117] "RemoveContainer" containerID="f6fdd19dac506f2d5d7eedc44ec2cda92a9381b73e91a45d949e4fcde4dfc283" Feb 02 15:19:20 crc kubenswrapper[4733]: E0202 15:19:20.078004 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6fdd19dac506f2d5d7eedc44ec2cda92a9381b73e91a45d949e4fcde4dfc283\": container with ID starting with f6fdd19dac506f2d5d7eedc44ec2cda92a9381b73e91a45d949e4fcde4dfc283 not found: ID does not exist" containerID="f6fdd19dac506f2d5d7eedc44ec2cda92a9381b73e91a45d949e4fcde4dfc283" Feb 02 15:19:20 crc kubenswrapper[4733]: I0202 15:19:20.078038 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6fdd19dac506f2d5d7eedc44ec2cda92a9381b73e91a45d949e4fcde4dfc283"} err="failed to get container status \"f6fdd19dac506f2d5d7eedc44ec2cda92a9381b73e91a45d949e4fcde4dfc283\": rpc error: code = NotFound desc = could not find container \"f6fdd19dac506f2d5d7eedc44ec2cda92a9381b73e91a45d949e4fcde4dfc283\": container with ID starting with f6fdd19dac506f2d5d7eedc44ec2cda92a9381b73e91a45d949e4fcde4dfc283 not found: ID does not exist" Feb 02 15:19:20 crc kubenswrapper[4733]: I0202 15:19:20.078076 4733 scope.go:117] "RemoveContainer" containerID="fd383d69b9b27d52fa8672ff6e0294a74876d676973bdc0ca95125fef5f2d085" Feb 02 15:19:20 crc kubenswrapper[4733]: E0202 15:19:20.079494 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd383d69b9b27d52fa8672ff6e0294a74876d676973bdc0ca95125fef5f2d085\": container with ID starting with fd383d69b9b27d52fa8672ff6e0294a74876d676973bdc0ca95125fef5f2d085 not found: ID does not exist" containerID="fd383d69b9b27d52fa8672ff6e0294a74876d676973bdc0ca95125fef5f2d085" Feb 02 15:19:20 crc kubenswrapper[4733]: I0202 15:19:20.079549 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd383d69b9b27d52fa8672ff6e0294a74876d676973bdc0ca95125fef5f2d085"} err="failed to get container status \"fd383d69b9b27d52fa8672ff6e0294a74876d676973bdc0ca95125fef5f2d085\": rpc error: code = NotFound desc = could not find container \"fd383d69b9b27d52fa8672ff6e0294a74876d676973bdc0ca95125fef5f2d085\": container with ID starting with fd383d69b9b27d52fa8672ff6e0294a74876d676973bdc0ca95125fef5f2d085 not found: ID does not exist" Feb 02 15:19:20 crc kubenswrapper[4733]: I0202 15:19:20.079580 4733 scope.go:117] "RemoveContainer" containerID="61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18" Feb 02 15:19:20 crc kubenswrapper[4733]: E0202 15:19:20.079941 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\": container with ID starting with 61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18 not found: ID does not exist" containerID="61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18" Feb 02 15:19:20 crc kubenswrapper[4733]: I0202 15:19:20.079964 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18"} err="failed to get container status \"61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\": rpc error: code = NotFound desc = could not find container \"61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18\": container with ID starting with 61717fcebfd7784aaea484098f29f4585f50cf39f079b57b733c73004b9f4c18 not found: ID does not exist" Feb 02 15:19:20 crc kubenswrapper[4733]: E0202 15:19:20.141499 4733 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.2:6443: connect: connection refused" interval="3.2s" Feb 02 15:19:21 crc kubenswrapper[4733]: I0202 15:19:21.265901 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Feb 02 15:19:21 crc kubenswrapper[4733]: E0202 15:19:21.334053 4733 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.2:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" volumeName="registry-storage" Feb 02 15:19:23 crc kubenswrapper[4733]: E0202 15:19:23.342570 4733 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.2:6443: connect: connection refused" interval="6.4s" Feb 02 15:19:27 crc kubenswrapper[4733]: I0202 15:19:27.260579 4733 status_manager.go:851] "Failed to get status for pod" podUID="272ea61e-a61b-40c8-8e4a-cdc9f3c95847" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.2:6443: connect: connection refused" Feb 02 15:19:29 crc kubenswrapper[4733]: E0202 15:19:29.744803 4733 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.2:6443: connect: connection refused" interval="7s" Feb 02 15:19:29 crc kubenswrapper[4733]: E0202 15:19:29.788276 4733 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.2:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.1890770e69673103 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-02 15:19:17.338222851 +0000 UTC m=+240.789684219,LastTimestamp:2026-02-02 15:19:17.338222851 +0000 UTC m=+240.789684219,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 02 15:19:30 crc kubenswrapper[4733]: I0202 15:19:30.254360 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 15:19:30 crc kubenswrapper[4733]: I0202 15:19:30.255515 4733 status_manager.go:851] "Failed to get status for pod" podUID="272ea61e-a61b-40c8-8e4a-cdc9f3c95847" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.2:6443: connect: connection refused" Feb 02 15:19:30 crc kubenswrapper[4733]: I0202 15:19:30.279102 4733 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c7082da2-f2e7-4b73-ba35-504e5cd4c920" Feb 02 15:19:30 crc kubenswrapper[4733]: I0202 15:19:30.279153 4733 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c7082da2-f2e7-4b73-ba35-504e5cd4c920" Feb 02 15:19:30 crc kubenswrapper[4733]: E0202 15:19:30.279792 4733 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.2:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 15:19:30 crc kubenswrapper[4733]: I0202 15:19:30.280414 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 15:19:30 crc kubenswrapper[4733]: W0202 15:19:30.313378 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-631abc5be6fc60f6d2e604d495cc1b9df1abaa973a38f7f9f36f24793697423a WatchSource:0}: Error finding container 631abc5be6fc60f6d2e604d495cc1b9df1abaa973a38f7f9f36f24793697423a: Status 404 returned error can't find the container with id 631abc5be6fc60f6d2e604d495cc1b9df1abaa973a38f7f9f36f24793697423a Feb 02 15:19:30 crc kubenswrapper[4733]: I0202 15:19:30.999344 4733 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="d9ae44de47981d0a43de400af177cde11a7db4709509cbc7d2c96c7c4eb038ee" exitCode=0 Feb 02 15:19:30 crc kubenswrapper[4733]: I0202 15:19:30.999410 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"d9ae44de47981d0a43de400af177cde11a7db4709509cbc7d2c96c7c4eb038ee"} Feb 02 15:19:30 crc kubenswrapper[4733]: I0202 15:19:30.999448 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"631abc5be6fc60f6d2e604d495cc1b9df1abaa973a38f7f9f36f24793697423a"} Feb 02 15:19:31 crc kubenswrapper[4733]: I0202 15:19:30.999915 4733 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c7082da2-f2e7-4b73-ba35-504e5cd4c920" Feb 02 15:19:31 crc kubenswrapper[4733]: I0202 15:19:30.999945 4733 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c7082da2-f2e7-4b73-ba35-504e5cd4c920" Feb 02 15:19:31 crc kubenswrapper[4733]: E0202 15:19:31.000920 4733 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.2:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 15:19:31 crc kubenswrapper[4733]: I0202 15:19:31.000933 4733 status_manager.go:851] "Failed to get status for pod" podUID="272ea61e-a61b-40c8-8e4a-cdc9f3c95847" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.2:6443: connect: connection refused" Feb 02 15:19:32 crc kubenswrapper[4733]: I0202 15:19:32.006826 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"a8b69c07899bae22662aea5e388413bb1c6e444cfe4049aa56a7ff1bc2e6c99b"} Feb 02 15:19:32 crc kubenswrapper[4733]: I0202 15:19:32.007115 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"f4548115e809d3431ff444971cddf7459893f933a31970830ec4fbead27af3fa"} Feb 02 15:19:32 crc kubenswrapper[4733]: I0202 15:19:32.007125 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"f837a699fdac8b3c86b30eacdcdbb3b5c1461103d8650f390afa68cf059da6c8"} Feb 02 15:19:32 crc kubenswrapper[4733]: I0202 15:19:32.009047 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Feb 02 15:19:32 crc kubenswrapper[4733]: I0202 15:19:32.009076 4733 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="7a3fd02d55a10647428a05a9f53a1429a9827059c523d5deab3c4581c8f1b933" exitCode=1 Feb 02 15:19:32 crc kubenswrapper[4733]: I0202 15:19:32.009095 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"7a3fd02d55a10647428a05a9f53a1429a9827059c523d5deab3c4581c8f1b933"} Feb 02 15:19:32 crc kubenswrapper[4733]: I0202 15:19:32.009538 4733 scope.go:117] "RemoveContainer" containerID="7a3fd02d55a10647428a05a9f53a1429a9827059c523d5deab3c4581c8f1b933" Feb 02 15:19:33 crc kubenswrapper[4733]: I0202 15:19:33.024117 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Feb 02 15:19:33 crc kubenswrapper[4733]: I0202 15:19:33.024427 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"9f7518e7686e9b4e8aa86845341e7f32d3682ea49f37d178c7d67e786c8ae891"} Feb 02 15:19:33 crc kubenswrapper[4733]: I0202 15:19:33.027313 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"feb6a87d39345997c589f92672de75e1b374c6f0de59032e2a511a631a4815f5"} Feb 02 15:19:33 crc kubenswrapper[4733]: I0202 15:19:33.027338 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"9039dbfe1c293f9597af0a4cf0841760c57631c41b983931c6f20e9c87116f95"} Feb 02 15:19:33 crc kubenswrapper[4733]: I0202 15:19:33.027536 4733 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c7082da2-f2e7-4b73-ba35-504e5cd4c920" Feb 02 15:19:33 crc kubenswrapper[4733]: I0202 15:19:33.027556 4733 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c7082da2-f2e7-4b73-ba35-504e5cd4c920" Feb 02 15:19:33 crc kubenswrapper[4733]: I0202 15:19:33.027751 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 15:19:33 crc kubenswrapper[4733]: I0202 15:19:33.959918 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 15:19:35 crc kubenswrapper[4733]: I0202 15:19:35.281413 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 15:19:35 crc kubenswrapper[4733]: I0202 15:19:35.281751 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 15:19:35 crc kubenswrapper[4733]: I0202 15:19:35.288035 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 15:19:38 crc kubenswrapper[4733]: I0202 15:19:38.035788 4733 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 15:19:39 crc kubenswrapper[4733]: I0202 15:19:39.087464 4733 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c7082da2-f2e7-4b73-ba35-504e5cd4c920" Feb 02 15:19:39 crc kubenswrapper[4733]: I0202 15:19:39.087787 4733 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c7082da2-f2e7-4b73-ba35-504e5cd4c920" Feb 02 15:19:39 crc kubenswrapper[4733]: I0202 15:19:39.091777 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 15:19:39 crc kubenswrapper[4733]: I0202 15:19:39.093571 4733 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="a2b37010-d792-4493-8b32-4aba4ae7e2a9" Feb 02 15:19:40 crc kubenswrapper[4733]: I0202 15:19:40.087067 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 15:19:40 crc kubenswrapper[4733]: I0202 15:19:40.095668 4733 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c7082da2-f2e7-4b73-ba35-504e5cd4c920" Feb 02 15:19:40 crc kubenswrapper[4733]: I0202 15:19:40.095886 4733 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c7082da2-f2e7-4b73-ba35-504e5cd4c920" Feb 02 15:19:40 crc kubenswrapper[4733]: I0202 15:19:40.096535 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 15:19:43 crc kubenswrapper[4733]: I0202 15:19:43.966651 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 02 15:19:47 crc kubenswrapper[4733]: I0202 15:19:47.274536 4733 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="a2b37010-d792-4493-8b32-4aba4ae7e2a9" Feb 02 15:19:48 crc kubenswrapper[4733]: I0202 15:19:48.667460 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 02 15:19:48 crc kubenswrapper[4733]: I0202 15:19:48.737567 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Feb 02 15:19:49 crc kubenswrapper[4733]: I0202 15:19:49.363804 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Feb 02 15:19:49 crc kubenswrapper[4733]: I0202 15:19:49.375513 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Feb 02 15:19:49 crc kubenswrapper[4733]: I0202 15:19:49.742360 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Feb 02 15:19:49 crc kubenswrapper[4733]: I0202 15:19:49.768829 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Feb 02 15:19:49 crc kubenswrapper[4733]: I0202 15:19:49.774018 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Feb 02 15:19:49 crc kubenswrapper[4733]: I0202 15:19:49.792003 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Feb 02 15:19:49 crc kubenswrapper[4733]: I0202 15:19:49.874074 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Feb 02 15:19:49 crc kubenswrapper[4733]: I0202 15:19:49.907913 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Feb 02 15:19:49 crc kubenswrapper[4733]: I0202 15:19:49.924920 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Feb 02 15:19:49 crc kubenswrapper[4733]: I0202 15:19:49.950764 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Feb 02 15:19:50 crc kubenswrapper[4733]: I0202 15:19:50.018154 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Feb 02 15:19:50 crc kubenswrapper[4733]: I0202 15:19:50.126830 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Feb 02 15:19:50 crc kubenswrapper[4733]: I0202 15:19:50.255024 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 02 15:19:50 crc kubenswrapper[4733]: I0202 15:19:50.263879 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Feb 02 15:19:50 crc kubenswrapper[4733]: I0202 15:19:50.389314 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Feb 02 15:19:50 crc kubenswrapper[4733]: I0202 15:19:50.554224 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Feb 02 15:19:50 crc kubenswrapper[4733]: I0202 15:19:50.655989 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Feb 02 15:19:50 crc kubenswrapper[4733]: I0202 15:19:50.774100 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Feb 02 15:19:50 crc kubenswrapper[4733]: I0202 15:19:50.805762 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Feb 02 15:19:50 crc kubenswrapper[4733]: I0202 15:19:50.856243 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Feb 02 15:19:50 crc kubenswrapper[4733]: I0202 15:19:50.892677 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Feb 02 15:19:50 crc kubenswrapper[4733]: I0202 15:19:50.990228 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Feb 02 15:19:51 crc kubenswrapper[4733]: I0202 15:19:51.038386 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Feb 02 15:19:51 crc kubenswrapper[4733]: I0202 15:19:51.064251 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Feb 02 15:19:51 crc kubenswrapper[4733]: I0202 15:19:51.072875 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Feb 02 15:19:51 crc kubenswrapper[4733]: I0202 15:19:51.073674 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 02 15:19:51 crc kubenswrapper[4733]: I0202 15:19:51.103140 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Feb 02 15:19:51 crc kubenswrapper[4733]: I0202 15:19:51.181519 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Feb 02 15:19:51 crc kubenswrapper[4733]: I0202 15:19:51.183576 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Feb 02 15:19:51 crc kubenswrapper[4733]: I0202 15:19:51.350107 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Feb 02 15:19:51 crc kubenswrapper[4733]: I0202 15:19:51.368393 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 02 15:19:51 crc kubenswrapper[4733]: I0202 15:19:51.529561 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Feb 02 15:19:51 crc kubenswrapper[4733]: I0202 15:19:51.531414 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Feb 02 15:19:51 crc kubenswrapper[4733]: I0202 15:19:51.709243 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Feb 02 15:19:51 crc kubenswrapper[4733]: I0202 15:19:51.770309 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Feb 02 15:19:51 crc kubenswrapper[4733]: I0202 15:19:51.804558 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Feb 02 15:19:51 crc kubenswrapper[4733]: I0202 15:19:51.820612 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Feb 02 15:19:51 crc kubenswrapper[4733]: I0202 15:19:51.872847 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Feb 02 15:19:51 crc kubenswrapper[4733]: I0202 15:19:51.940988 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Feb 02 15:19:51 crc kubenswrapper[4733]: I0202 15:19:51.966599 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Feb 02 15:19:52 crc kubenswrapper[4733]: I0202 15:19:52.021288 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Feb 02 15:19:52 crc kubenswrapper[4733]: I0202 15:19:52.172121 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Feb 02 15:19:52 crc kubenswrapper[4733]: I0202 15:19:52.315545 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Feb 02 15:19:52 crc kubenswrapper[4733]: I0202 15:19:52.322418 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 02 15:19:52 crc kubenswrapper[4733]: I0202 15:19:52.346769 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Feb 02 15:19:52 crc kubenswrapper[4733]: I0202 15:19:52.379231 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Feb 02 15:19:52 crc kubenswrapper[4733]: I0202 15:19:52.389137 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 02 15:19:52 crc kubenswrapper[4733]: I0202 15:19:52.416666 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Feb 02 15:19:52 crc kubenswrapper[4733]: I0202 15:19:52.695365 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Feb 02 15:19:52 crc kubenswrapper[4733]: I0202 15:19:52.721576 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Feb 02 15:19:52 crc kubenswrapper[4733]: I0202 15:19:52.733779 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Feb 02 15:19:52 crc kubenswrapper[4733]: I0202 15:19:52.840522 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Feb 02 15:19:52 crc kubenswrapper[4733]: I0202 15:19:52.860779 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 02 15:19:52 crc kubenswrapper[4733]: I0202 15:19:52.909691 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Feb 02 15:19:52 crc kubenswrapper[4733]: I0202 15:19:52.934015 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Feb 02 15:19:52 crc kubenswrapper[4733]: I0202 15:19:52.961661 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Feb 02 15:19:52 crc kubenswrapper[4733]: I0202 15:19:52.974458 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Feb 02 15:19:52 crc kubenswrapper[4733]: I0202 15:19:52.999697 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Feb 02 15:19:53 crc kubenswrapper[4733]: I0202 15:19:53.004041 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Feb 02 15:19:53 crc kubenswrapper[4733]: I0202 15:19:53.044296 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Feb 02 15:19:53 crc kubenswrapper[4733]: I0202 15:19:53.069557 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Feb 02 15:19:53 crc kubenswrapper[4733]: I0202 15:19:53.094768 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Feb 02 15:19:53 crc kubenswrapper[4733]: I0202 15:19:53.153210 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Feb 02 15:19:53 crc kubenswrapper[4733]: I0202 15:19:53.186409 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Feb 02 15:19:53 crc kubenswrapper[4733]: I0202 15:19:53.275279 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Feb 02 15:19:53 crc kubenswrapper[4733]: I0202 15:19:53.290422 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Feb 02 15:19:53 crc kubenswrapper[4733]: I0202 15:19:53.381754 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Feb 02 15:19:53 crc kubenswrapper[4733]: I0202 15:19:53.428415 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Feb 02 15:19:53 crc kubenswrapper[4733]: I0202 15:19:53.438986 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Feb 02 15:19:53 crc kubenswrapper[4733]: I0202 15:19:53.492759 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Feb 02 15:19:53 crc kubenswrapper[4733]: I0202 15:19:53.499667 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Feb 02 15:19:53 crc kubenswrapper[4733]: I0202 15:19:53.508438 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 02 15:19:53 crc kubenswrapper[4733]: I0202 15:19:53.511513 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 02 15:19:53 crc kubenswrapper[4733]: I0202 15:19:53.554560 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Feb 02 15:19:53 crc kubenswrapper[4733]: I0202 15:19:53.635925 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 02 15:19:53 crc kubenswrapper[4733]: I0202 15:19:53.691769 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Feb 02 15:19:53 crc kubenswrapper[4733]: I0202 15:19:53.725429 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Feb 02 15:19:53 crc kubenswrapper[4733]: I0202 15:19:53.826911 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Feb 02 15:19:53 crc kubenswrapper[4733]: I0202 15:19:53.855830 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Feb 02 15:19:53 crc kubenswrapper[4733]: I0202 15:19:53.946503 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Feb 02 15:19:54 crc kubenswrapper[4733]: I0202 15:19:54.004220 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 02 15:19:54 crc kubenswrapper[4733]: I0202 15:19:54.064943 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Feb 02 15:19:54 crc kubenswrapper[4733]: I0202 15:19:54.077825 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Feb 02 15:19:54 crc kubenswrapper[4733]: I0202 15:19:54.158863 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Feb 02 15:19:54 crc kubenswrapper[4733]: I0202 15:19:54.189616 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Feb 02 15:19:54 crc kubenswrapper[4733]: I0202 15:19:54.206477 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 02 15:19:54 crc kubenswrapper[4733]: I0202 15:19:54.233450 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 02 15:19:54 crc kubenswrapper[4733]: I0202 15:19:54.374155 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 02 15:19:54 crc kubenswrapper[4733]: I0202 15:19:54.381077 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Feb 02 15:19:54 crc kubenswrapper[4733]: I0202 15:19:54.392541 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Feb 02 15:19:54 crc kubenswrapper[4733]: I0202 15:19:54.462694 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Feb 02 15:19:54 crc kubenswrapper[4733]: I0202 15:19:54.502125 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Feb 02 15:19:54 crc kubenswrapper[4733]: I0202 15:19:54.509340 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Feb 02 15:19:54 crc kubenswrapper[4733]: I0202 15:19:54.525660 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Feb 02 15:19:54 crc kubenswrapper[4733]: I0202 15:19:54.630340 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Feb 02 15:19:54 crc kubenswrapper[4733]: I0202 15:19:54.650973 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Feb 02 15:19:54 crc kubenswrapper[4733]: I0202 15:19:54.716240 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Feb 02 15:19:54 crc kubenswrapper[4733]: I0202 15:19:54.887110 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Feb 02 15:19:54 crc kubenswrapper[4733]: I0202 15:19:54.895284 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Feb 02 15:19:54 crc kubenswrapper[4733]: I0202 15:19:54.937746 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Feb 02 15:19:54 crc kubenswrapper[4733]: I0202 15:19:54.954022 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Feb 02 15:19:55 crc kubenswrapper[4733]: I0202 15:19:55.069815 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Feb 02 15:19:55 crc kubenswrapper[4733]: I0202 15:19:55.306996 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Feb 02 15:19:55 crc kubenswrapper[4733]: I0202 15:19:55.411607 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 02 15:19:55 crc kubenswrapper[4733]: I0202 15:19:55.461260 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Feb 02 15:19:55 crc kubenswrapper[4733]: I0202 15:19:55.465389 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Feb 02 15:19:55 crc kubenswrapper[4733]: I0202 15:19:55.561486 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Feb 02 15:19:55 crc kubenswrapper[4733]: I0202 15:19:55.659781 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Feb 02 15:19:55 crc kubenswrapper[4733]: I0202 15:19:55.838838 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Feb 02 15:19:55 crc kubenswrapper[4733]: I0202 15:19:55.937579 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Feb 02 15:19:55 crc kubenswrapper[4733]: I0202 15:19:55.939048 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Feb 02 15:19:55 crc kubenswrapper[4733]: I0202 15:19:55.951281 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 02 15:19:55 crc kubenswrapper[4733]: I0202 15:19:55.969957 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Feb 02 15:19:55 crc kubenswrapper[4733]: I0202 15:19:55.974098 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Feb 02 15:19:56 crc kubenswrapper[4733]: I0202 15:19:56.038441 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Feb 02 15:19:56 crc kubenswrapper[4733]: I0202 15:19:56.144621 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Feb 02 15:19:56 crc kubenswrapper[4733]: I0202 15:19:56.152263 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Feb 02 15:19:56 crc kubenswrapper[4733]: I0202 15:19:56.174891 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 02 15:19:56 crc kubenswrapper[4733]: I0202 15:19:56.220412 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Feb 02 15:19:56 crc kubenswrapper[4733]: I0202 15:19:56.225432 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Feb 02 15:19:56 crc kubenswrapper[4733]: I0202 15:19:56.251708 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Feb 02 15:19:56 crc kubenswrapper[4733]: I0202 15:19:56.263056 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Feb 02 15:19:56 crc kubenswrapper[4733]: I0202 15:19:56.272410 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Feb 02 15:19:56 crc kubenswrapper[4733]: I0202 15:19:56.278756 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Feb 02 15:19:56 crc kubenswrapper[4733]: I0202 15:19:56.352520 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Feb 02 15:19:56 crc kubenswrapper[4733]: I0202 15:19:56.384045 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Feb 02 15:19:56 crc kubenswrapper[4733]: I0202 15:19:56.439241 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Feb 02 15:19:56 crc kubenswrapper[4733]: I0202 15:19:56.498977 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Feb 02 15:19:56 crc kubenswrapper[4733]: I0202 15:19:56.518814 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Feb 02 15:19:56 crc kubenswrapper[4733]: I0202 15:19:56.592829 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Feb 02 15:19:56 crc kubenswrapper[4733]: I0202 15:19:56.728651 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Feb 02 15:19:56 crc kubenswrapper[4733]: I0202 15:19:56.777644 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 02 15:19:56 crc kubenswrapper[4733]: I0202 15:19:56.803025 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Feb 02 15:19:56 crc kubenswrapper[4733]: I0202 15:19:56.819775 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Feb 02 15:19:57 crc kubenswrapper[4733]: I0202 15:19:57.006251 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Feb 02 15:19:57 crc kubenswrapper[4733]: I0202 15:19:57.092998 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Feb 02 15:19:57 crc kubenswrapper[4733]: I0202 15:19:57.161392 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Feb 02 15:19:57 crc kubenswrapper[4733]: I0202 15:19:57.174731 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Feb 02 15:19:57 crc kubenswrapper[4733]: I0202 15:19:57.179010 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Feb 02 15:19:57 crc kubenswrapper[4733]: I0202 15:19:57.197194 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Feb 02 15:19:57 crc kubenswrapper[4733]: I0202 15:19:57.225592 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Feb 02 15:19:57 crc kubenswrapper[4733]: I0202 15:19:57.252331 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Feb 02 15:19:57 crc kubenswrapper[4733]: I0202 15:19:57.269863 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Feb 02 15:19:57 crc kubenswrapper[4733]: I0202 15:19:57.279156 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Feb 02 15:19:57 crc kubenswrapper[4733]: I0202 15:19:57.322018 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Feb 02 15:19:57 crc kubenswrapper[4733]: I0202 15:19:57.332882 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 02 15:19:57 crc kubenswrapper[4733]: I0202 15:19:57.351651 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Feb 02 15:19:57 crc kubenswrapper[4733]: I0202 15:19:57.415118 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Feb 02 15:19:57 crc kubenswrapper[4733]: I0202 15:19:57.417037 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Feb 02 15:19:57 crc kubenswrapper[4733]: I0202 15:19:57.474855 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Feb 02 15:19:57 crc kubenswrapper[4733]: I0202 15:19:57.519343 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Feb 02 15:19:57 crc kubenswrapper[4733]: I0202 15:19:57.531991 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Feb 02 15:19:57 crc kubenswrapper[4733]: I0202 15:19:57.653342 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Feb 02 15:19:57 crc kubenswrapper[4733]: I0202 15:19:57.787085 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Feb 02 15:19:57 crc kubenswrapper[4733]: I0202 15:19:57.795100 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Feb 02 15:19:57 crc kubenswrapper[4733]: I0202 15:19:57.828106 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Feb 02 15:19:57 crc kubenswrapper[4733]: I0202 15:19:57.944516 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Feb 02 15:19:57 crc kubenswrapper[4733]: I0202 15:19:57.956930 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 02 15:19:58 crc kubenswrapper[4733]: I0202 15:19:58.002120 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Feb 02 15:19:58 crc kubenswrapper[4733]: I0202 15:19:58.221643 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Feb 02 15:19:58 crc kubenswrapper[4733]: I0202 15:19:58.264080 4733 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 02 15:19:58 crc kubenswrapper[4733]: I0202 15:19:58.306826 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Feb 02 15:19:58 crc kubenswrapper[4733]: I0202 15:19:58.307733 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Feb 02 15:19:58 crc kubenswrapper[4733]: I0202 15:19:58.345035 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Feb 02 15:19:58 crc kubenswrapper[4733]: I0202 15:19:58.433139 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Feb 02 15:19:58 crc kubenswrapper[4733]: I0202 15:19:58.554130 4733 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Feb 02 15:19:58 crc kubenswrapper[4733]: I0202 15:19:58.601906 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Feb 02 15:19:58 crc kubenswrapper[4733]: I0202 15:19:58.614356 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Feb 02 15:19:58 crc kubenswrapper[4733]: I0202 15:19:58.633861 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Feb 02 15:19:58 crc kubenswrapper[4733]: I0202 15:19:58.658882 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Feb 02 15:19:58 crc kubenswrapper[4733]: I0202 15:19:58.663616 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Feb 02 15:19:58 crc kubenswrapper[4733]: I0202 15:19:58.834123 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Feb 02 15:19:58 crc kubenswrapper[4733]: I0202 15:19:58.836352 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 02 15:19:58 crc kubenswrapper[4733]: I0202 15:19:58.856514 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Feb 02 15:19:58 crc kubenswrapper[4733]: I0202 15:19:58.862585 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Feb 02 15:19:58 crc kubenswrapper[4733]: I0202 15:19:58.878418 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Feb 02 15:19:58 crc kubenswrapper[4733]: I0202 15:19:58.919929 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Feb 02 15:19:58 crc kubenswrapper[4733]: I0202 15:19:58.983152 4733 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Feb 02 15:19:58 crc kubenswrapper[4733]: I0202 15:19:58.993433 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 02 15:19:58 crc kubenswrapper[4733]: I0202 15:19:58.993487 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 02 15:19:59 crc kubenswrapper[4733]: I0202 15:19:59.002249 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 02 15:19:59 crc kubenswrapper[4733]: I0202 15:19:59.017120 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=21.017100872 podStartE2EDuration="21.017100872s" podCreationTimestamp="2026-02-02 15:19:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:19:59.013711491 +0000 UTC m=+282.465172879" watchObservedRunningTime="2026-02-02 15:19:59.017100872 +0000 UTC m=+282.468562240" Feb 02 15:19:59 crc kubenswrapper[4733]: I0202 15:19:59.063577 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Feb 02 15:19:59 crc kubenswrapper[4733]: I0202 15:19:59.111668 4733 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Feb 02 15:19:59 crc kubenswrapper[4733]: I0202 15:19:59.121427 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 02 15:19:59 crc kubenswrapper[4733]: I0202 15:19:59.127532 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Feb 02 15:19:59 crc kubenswrapper[4733]: I0202 15:19:59.189888 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Feb 02 15:19:59 crc kubenswrapper[4733]: I0202 15:19:59.331754 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 02 15:19:59 crc kubenswrapper[4733]: I0202 15:19:59.477112 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Feb 02 15:19:59 crc kubenswrapper[4733]: I0202 15:19:59.484629 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Feb 02 15:19:59 crc kubenswrapper[4733]: I0202 15:19:59.534452 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Feb 02 15:19:59 crc kubenswrapper[4733]: I0202 15:19:59.579102 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Feb 02 15:19:59 crc kubenswrapper[4733]: I0202 15:19:59.614740 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Feb 02 15:19:59 crc kubenswrapper[4733]: I0202 15:19:59.663469 4733 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Feb 02 15:19:59 crc kubenswrapper[4733]: I0202 15:19:59.751294 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Feb 02 15:19:59 crc kubenswrapper[4733]: I0202 15:19:59.807067 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 02 15:19:59 crc kubenswrapper[4733]: I0202 15:19:59.809177 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Feb 02 15:19:59 crc kubenswrapper[4733]: I0202 15:19:59.815999 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Feb 02 15:19:59 crc kubenswrapper[4733]: I0202 15:19:59.965015 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 02 15:19:59 crc kubenswrapper[4733]: I0202 15:19:59.981759 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 02 15:19:59 crc kubenswrapper[4733]: I0202 15:19:59.984541 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 02 15:20:00 crc kubenswrapper[4733]: I0202 15:20:00.002778 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Feb 02 15:20:00 crc kubenswrapper[4733]: I0202 15:20:00.083913 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Feb 02 15:20:00 crc kubenswrapper[4733]: I0202 15:20:00.135352 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Feb 02 15:20:00 crc kubenswrapper[4733]: I0202 15:20:00.255047 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Feb 02 15:20:00 crc kubenswrapper[4733]: I0202 15:20:00.311624 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 02 15:20:00 crc kubenswrapper[4733]: I0202 15:20:00.386598 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Feb 02 15:20:00 crc kubenswrapper[4733]: I0202 15:20:00.482276 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Feb 02 15:20:00 crc kubenswrapper[4733]: I0202 15:20:00.483904 4733 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 02 15:20:00 crc kubenswrapper[4733]: I0202 15:20:00.484283 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://c5c8bdc71a33879c2f3adefae66e3670d35ca4234942866cf4c153aee739ccfa" gracePeriod=5 Feb 02 15:20:00 crc kubenswrapper[4733]: I0202 15:20:00.502669 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 02 15:20:00 crc kubenswrapper[4733]: I0202 15:20:00.522896 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Feb 02 15:20:00 crc kubenswrapper[4733]: I0202 15:20:00.567754 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Feb 02 15:20:00 crc kubenswrapper[4733]: I0202 15:20:00.617412 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Feb 02 15:20:00 crc kubenswrapper[4733]: I0202 15:20:00.735693 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Feb 02 15:20:00 crc kubenswrapper[4733]: I0202 15:20:00.739115 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Feb 02 15:20:00 crc kubenswrapper[4733]: I0202 15:20:00.767752 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Feb 02 15:20:00 crc kubenswrapper[4733]: I0202 15:20:00.788764 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Feb 02 15:20:00 crc kubenswrapper[4733]: I0202 15:20:00.914010 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Feb 02 15:20:01 crc kubenswrapper[4733]: I0202 15:20:01.087784 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Feb 02 15:20:01 crc kubenswrapper[4733]: I0202 15:20:01.124683 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Feb 02 15:20:01 crc kubenswrapper[4733]: I0202 15:20:01.291445 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Feb 02 15:20:01 crc kubenswrapper[4733]: I0202 15:20:01.313795 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Feb 02 15:20:01 crc kubenswrapper[4733]: I0202 15:20:01.392893 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 02 15:20:01 crc kubenswrapper[4733]: I0202 15:20:01.434025 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Feb 02 15:20:01 crc kubenswrapper[4733]: I0202 15:20:01.479057 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Feb 02 15:20:01 crc kubenswrapper[4733]: I0202 15:20:01.484309 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Feb 02 15:20:01 crc kubenswrapper[4733]: I0202 15:20:01.592465 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Feb 02 15:20:01 crc kubenswrapper[4733]: I0202 15:20:01.643529 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Feb 02 15:20:01 crc kubenswrapper[4733]: I0202 15:20:01.657634 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Feb 02 15:20:01 crc kubenswrapper[4733]: I0202 15:20:01.657903 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Feb 02 15:20:01 crc kubenswrapper[4733]: I0202 15:20:01.664012 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Feb 02 15:20:01 crc kubenswrapper[4733]: I0202 15:20:01.810993 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Feb 02 15:20:01 crc kubenswrapper[4733]: I0202 15:20:01.874131 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Feb 02 15:20:01 crc kubenswrapper[4733]: I0202 15:20:01.888914 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Feb 02 15:20:02 crc kubenswrapper[4733]: I0202 15:20:02.156349 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Feb 02 15:20:02 crc kubenswrapper[4733]: I0202 15:20:02.339747 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 02 15:20:02 crc kubenswrapper[4733]: I0202 15:20:02.359540 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Feb 02 15:20:02 crc kubenswrapper[4733]: I0202 15:20:02.367594 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Feb 02 15:20:02 crc kubenswrapper[4733]: I0202 15:20:02.676342 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Feb 02 15:20:02 crc kubenswrapper[4733]: I0202 15:20:02.802278 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 02 15:20:03 crc kubenswrapper[4733]: I0202 15:20:03.158692 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Feb 02 15:20:03 crc kubenswrapper[4733]: I0202 15:20:03.160351 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Feb 02 15:20:03 crc kubenswrapper[4733]: I0202 15:20:03.188448 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Feb 02 15:20:03 crc kubenswrapper[4733]: I0202 15:20:03.269818 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Feb 02 15:20:03 crc kubenswrapper[4733]: I0202 15:20:03.372480 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Feb 02 15:20:03 crc kubenswrapper[4733]: I0202 15:20:03.461441 4733 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Feb 02 15:20:03 crc kubenswrapper[4733]: I0202 15:20:03.549560 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Feb 02 15:20:03 crc kubenswrapper[4733]: I0202 15:20:03.587229 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Feb 02 15:20:04 crc kubenswrapper[4733]: I0202 15:20:04.003781 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Feb 02 15:20:04 crc kubenswrapper[4733]: I0202 15:20:04.318041 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Feb 02 15:20:04 crc kubenswrapper[4733]: I0202 15:20:04.436718 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Feb 02 15:20:04 crc kubenswrapper[4733]: I0202 15:20:04.437048 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Feb 02 15:20:04 crc kubenswrapper[4733]: I0202 15:20:04.489475 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Feb 02 15:20:04 crc kubenswrapper[4733]: I0202 15:20:04.565410 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Feb 02 15:20:05 crc kubenswrapper[4733]: I0202 15:20:05.370770 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Feb 02 15:20:06 crc kubenswrapper[4733]: I0202 15:20:06.079933 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 02 15:20:06 crc kubenswrapper[4733]: I0202 15:20:06.080661 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 15:20:06 crc kubenswrapper[4733]: I0202 15:20:06.124659 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 02 15:20:06 crc kubenswrapper[4733]: I0202 15:20:06.124763 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 02 15:20:06 crc kubenswrapper[4733]: I0202 15:20:06.124815 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 02 15:20:06 crc kubenswrapper[4733]: I0202 15:20:06.124860 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 02 15:20:06 crc kubenswrapper[4733]: I0202 15:20:06.124889 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 02 15:20:06 crc kubenswrapper[4733]: I0202 15:20:06.124954 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 15:20:06 crc kubenswrapper[4733]: I0202 15:20:06.125020 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 15:20:06 crc kubenswrapper[4733]: I0202 15:20:06.125064 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 15:20:06 crc kubenswrapper[4733]: I0202 15:20:06.125090 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 15:20:06 crc kubenswrapper[4733]: I0202 15:20:06.125260 4733 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Feb 02 15:20:06 crc kubenswrapper[4733]: I0202 15:20:06.125285 4733 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Feb 02 15:20:06 crc kubenswrapper[4733]: I0202 15:20:06.125303 4733 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 02 15:20:06 crc kubenswrapper[4733]: I0202 15:20:06.125320 4733 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Feb 02 15:20:06 crc kubenswrapper[4733]: I0202 15:20:06.136359 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 15:20:06 crc kubenswrapper[4733]: I0202 15:20:06.226855 4733 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 02 15:20:06 crc kubenswrapper[4733]: I0202 15:20:06.288672 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 02 15:20:06 crc kubenswrapper[4733]: I0202 15:20:06.288727 4733 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="c5c8bdc71a33879c2f3adefae66e3670d35ca4234942866cf4c153aee739ccfa" exitCode=137 Feb 02 15:20:06 crc kubenswrapper[4733]: I0202 15:20:06.288772 4733 scope.go:117] "RemoveContainer" containerID="c5c8bdc71a33879c2f3adefae66e3670d35ca4234942866cf4c153aee739ccfa" Feb 02 15:20:06 crc kubenswrapper[4733]: I0202 15:20:06.288906 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 02 15:20:06 crc kubenswrapper[4733]: I0202 15:20:06.318375 4733 scope.go:117] "RemoveContainer" containerID="c5c8bdc71a33879c2f3adefae66e3670d35ca4234942866cf4c153aee739ccfa" Feb 02 15:20:06 crc kubenswrapper[4733]: E0202 15:20:06.318890 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5c8bdc71a33879c2f3adefae66e3670d35ca4234942866cf4c153aee739ccfa\": container with ID starting with c5c8bdc71a33879c2f3adefae66e3670d35ca4234942866cf4c153aee739ccfa not found: ID does not exist" containerID="c5c8bdc71a33879c2f3adefae66e3670d35ca4234942866cf4c153aee739ccfa" Feb 02 15:20:06 crc kubenswrapper[4733]: I0202 15:20:06.318927 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5c8bdc71a33879c2f3adefae66e3670d35ca4234942866cf4c153aee739ccfa"} err="failed to get container status \"c5c8bdc71a33879c2f3adefae66e3670d35ca4234942866cf4c153aee739ccfa\": rpc error: code = NotFound desc = could not find container \"c5c8bdc71a33879c2f3adefae66e3670d35ca4234942866cf4c153aee739ccfa\": container with ID starting with c5c8bdc71a33879c2f3adefae66e3670d35ca4234942866cf4c153aee739ccfa not found: ID does not exist" Feb 02 15:20:07 crc kubenswrapper[4733]: I0202 15:20:07.265712 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Feb 02 15:20:17 crc kubenswrapper[4733]: I0202 15:20:17.038757 4733 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Feb 02 15:20:32 crc kubenswrapper[4733]: I0202 15:20:32.519827 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-m62b2"] Feb 02 15:20:32 crc kubenswrapper[4733]: I0202 15:20:32.520497 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-m62b2" podUID="7718b7b0-f1a7-4b87-8ebe-345e9100463d" containerName="controller-manager" containerID="cri-o://e1d6f6e0975a0529556931f0c74afc5d70e93d838245956f2238a399298a6400" gracePeriod=30 Feb 02 15:20:32 crc kubenswrapper[4733]: I0202 15:20:32.531407 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-98h6l"] Feb 02 15:20:32 crc kubenswrapper[4733]: I0202 15:20:32.531663 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-98h6l" podUID="f7829943-cabc-49c5-a547-62124e38d49e" containerName="route-controller-manager" containerID="cri-o://918ae4dd606f38b41b783f31ac022be12bb0fe2af0666b1799e8dcec1e790b74" gracePeriod=30 Feb 02 15:20:35 crc kubenswrapper[4733]: I0202 15:20:35.479609 4733 generic.go:334] "Generic (PLEG): container finished" podID="7718b7b0-f1a7-4b87-8ebe-345e9100463d" containerID="e1d6f6e0975a0529556931f0c74afc5d70e93d838245956f2238a399298a6400" exitCode=0 Feb 02 15:20:35 crc kubenswrapper[4733]: I0202 15:20:35.479695 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-m62b2" event={"ID":"7718b7b0-f1a7-4b87-8ebe-345e9100463d","Type":"ContainerDied","Data":"e1d6f6e0975a0529556931f0c74afc5d70e93d838245956f2238a399298a6400"} Feb 02 15:20:35 crc kubenswrapper[4733]: I0202 15:20:35.482652 4733 generic.go:334] "Generic (PLEG): container finished" podID="f7829943-cabc-49c5-a547-62124e38d49e" containerID="918ae4dd606f38b41b783f31ac022be12bb0fe2af0666b1799e8dcec1e790b74" exitCode=0 Feb 02 15:20:35 crc kubenswrapper[4733]: I0202 15:20:35.482706 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-98h6l" event={"ID":"f7829943-cabc-49c5-a547-62124e38d49e","Type":"ContainerDied","Data":"918ae4dd606f38b41b783f31ac022be12bb0fe2af0666b1799e8dcec1e790b74"} Feb 02 15:20:35 crc kubenswrapper[4733]: I0202 15:20:35.836523 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-m62b2" Feb 02 15:20:35 crc kubenswrapper[4733]: I0202 15:20:35.842701 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-98h6l" Feb 02 15:20:35 crc kubenswrapper[4733]: I0202 15:20:35.864582 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-86cb48757f-sfmxz"] Feb 02 15:20:35 crc kubenswrapper[4733]: E0202 15:20:35.864764 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 02 15:20:35 crc kubenswrapper[4733]: I0202 15:20:35.864774 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 02 15:20:35 crc kubenswrapper[4733]: E0202 15:20:35.864786 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7829943-cabc-49c5-a547-62124e38d49e" containerName="route-controller-manager" Feb 02 15:20:35 crc kubenswrapper[4733]: I0202 15:20:35.864792 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7829943-cabc-49c5-a547-62124e38d49e" containerName="route-controller-manager" Feb 02 15:20:35 crc kubenswrapper[4733]: E0202 15:20:35.864802 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="272ea61e-a61b-40c8-8e4a-cdc9f3c95847" containerName="installer" Feb 02 15:20:35 crc kubenswrapper[4733]: I0202 15:20:35.864808 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="272ea61e-a61b-40c8-8e4a-cdc9f3c95847" containerName="installer" Feb 02 15:20:35 crc kubenswrapper[4733]: E0202 15:20:35.864815 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7718b7b0-f1a7-4b87-8ebe-345e9100463d" containerName="controller-manager" Feb 02 15:20:35 crc kubenswrapper[4733]: I0202 15:20:35.864820 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="7718b7b0-f1a7-4b87-8ebe-345e9100463d" containerName="controller-manager" Feb 02 15:20:35 crc kubenswrapper[4733]: I0202 15:20:35.864929 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7829943-cabc-49c5-a547-62124e38d49e" containerName="route-controller-manager" Feb 02 15:20:35 crc kubenswrapper[4733]: I0202 15:20:35.864939 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 02 15:20:35 crc kubenswrapper[4733]: I0202 15:20:35.864951 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="7718b7b0-f1a7-4b87-8ebe-345e9100463d" containerName="controller-manager" Feb 02 15:20:35 crc kubenswrapper[4733]: I0202 15:20:35.864957 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="272ea61e-a61b-40c8-8e4a-cdc9f3c95847" containerName="installer" Feb 02 15:20:35 crc kubenswrapper[4733]: I0202 15:20:35.865436 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-86cb48757f-sfmxz" Feb 02 15:20:35 crc kubenswrapper[4733]: I0202 15:20:35.881864 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-86cb48757f-sfmxz"] Feb 02 15:20:35 crc kubenswrapper[4733]: I0202 15:20:35.949532 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7718b7b0-f1a7-4b87-8ebe-345e9100463d-client-ca\") pod \"7718b7b0-f1a7-4b87-8ebe-345e9100463d\" (UID: \"7718b7b0-f1a7-4b87-8ebe-345e9100463d\") " Feb 02 15:20:35 crc kubenswrapper[4733]: I0202 15:20:35.949594 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7718b7b0-f1a7-4b87-8ebe-345e9100463d-serving-cert\") pod \"7718b7b0-f1a7-4b87-8ebe-345e9100463d\" (UID: \"7718b7b0-f1a7-4b87-8ebe-345e9100463d\") " Feb 02 15:20:35 crc kubenswrapper[4733]: I0202 15:20:35.949612 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f7829943-cabc-49c5-a547-62124e38d49e-client-ca\") pod \"f7829943-cabc-49c5-a547-62124e38d49e\" (UID: \"f7829943-cabc-49c5-a547-62124e38d49e\") " Feb 02 15:20:35 crc kubenswrapper[4733]: I0202 15:20:35.949637 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7829943-cabc-49c5-a547-62124e38d49e-config\") pod \"f7829943-cabc-49c5-a547-62124e38d49e\" (UID: \"f7829943-cabc-49c5-a547-62124e38d49e\") " Feb 02 15:20:35 crc kubenswrapper[4733]: I0202 15:20:35.949671 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5f8gt\" (UniqueName: \"kubernetes.io/projected/f7829943-cabc-49c5-a547-62124e38d49e-kube-api-access-5f8gt\") pod \"f7829943-cabc-49c5-a547-62124e38d49e\" (UID: \"f7829943-cabc-49c5-a547-62124e38d49e\") " Feb 02 15:20:35 crc kubenswrapper[4733]: I0202 15:20:35.949710 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7718b7b0-f1a7-4b87-8ebe-345e9100463d-config\") pod \"7718b7b0-f1a7-4b87-8ebe-345e9100463d\" (UID: \"7718b7b0-f1a7-4b87-8ebe-345e9100463d\") " Feb 02 15:20:35 crc kubenswrapper[4733]: I0202 15:20:35.949740 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vfvl2\" (UniqueName: \"kubernetes.io/projected/7718b7b0-f1a7-4b87-8ebe-345e9100463d-kube-api-access-vfvl2\") pod \"7718b7b0-f1a7-4b87-8ebe-345e9100463d\" (UID: \"7718b7b0-f1a7-4b87-8ebe-345e9100463d\") " Feb 02 15:20:35 crc kubenswrapper[4733]: I0202 15:20:35.949783 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7718b7b0-f1a7-4b87-8ebe-345e9100463d-proxy-ca-bundles\") pod \"7718b7b0-f1a7-4b87-8ebe-345e9100463d\" (UID: \"7718b7b0-f1a7-4b87-8ebe-345e9100463d\") " Feb 02 15:20:35 crc kubenswrapper[4733]: I0202 15:20:35.949825 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7829943-cabc-49c5-a547-62124e38d49e-serving-cert\") pod \"f7829943-cabc-49c5-a547-62124e38d49e\" (UID: \"f7829943-cabc-49c5-a547-62124e38d49e\") " Feb 02 15:20:35 crc kubenswrapper[4733]: I0202 15:20:35.950565 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2c0443cb-4d5e-4389-a935-1821f94f97a3-client-ca\") pod \"controller-manager-86cb48757f-sfmxz\" (UID: \"2c0443cb-4d5e-4389-a935-1821f94f97a3\") " pod="openshift-controller-manager/controller-manager-86cb48757f-sfmxz" Feb 02 15:20:35 crc kubenswrapper[4733]: I0202 15:20:35.950599 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2p6h\" (UniqueName: \"kubernetes.io/projected/2c0443cb-4d5e-4389-a935-1821f94f97a3-kube-api-access-r2p6h\") pod \"controller-manager-86cb48757f-sfmxz\" (UID: \"2c0443cb-4d5e-4389-a935-1821f94f97a3\") " pod="openshift-controller-manager/controller-manager-86cb48757f-sfmxz" Feb 02 15:20:35 crc kubenswrapper[4733]: I0202 15:20:35.950648 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2c0443cb-4d5e-4389-a935-1821f94f97a3-proxy-ca-bundles\") pod \"controller-manager-86cb48757f-sfmxz\" (UID: \"2c0443cb-4d5e-4389-a935-1821f94f97a3\") " pod="openshift-controller-manager/controller-manager-86cb48757f-sfmxz" Feb 02 15:20:35 crc kubenswrapper[4733]: I0202 15:20:35.950819 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f7829943-cabc-49c5-a547-62124e38d49e-client-ca" (OuterVolumeSpecName: "client-ca") pod "f7829943-cabc-49c5-a547-62124e38d49e" (UID: "f7829943-cabc-49c5-a547-62124e38d49e"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:20:35 crc kubenswrapper[4733]: I0202 15:20:35.950890 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7718b7b0-f1a7-4b87-8ebe-345e9100463d-client-ca" (OuterVolumeSpecName: "client-ca") pod "7718b7b0-f1a7-4b87-8ebe-345e9100463d" (UID: "7718b7b0-f1a7-4b87-8ebe-345e9100463d"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:20:35 crc kubenswrapper[4733]: I0202 15:20:35.950892 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c0443cb-4d5e-4389-a935-1821f94f97a3-config\") pod \"controller-manager-86cb48757f-sfmxz\" (UID: \"2c0443cb-4d5e-4389-a935-1821f94f97a3\") " pod="openshift-controller-manager/controller-manager-86cb48757f-sfmxz" Feb 02 15:20:35 crc kubenswrapper[4733]: I0202 15:20:35.950943 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f7829943-cabc-49c5-a547-62124e38d49e-config" (OuterVolumeSpecName: "config") pod "f7829943-cabc-49c5-a547-62124e38d49e" (UID: "f7829943-cabc-49c5-a547-62124e38d49e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:20:35 crc kubenswrapper[4733]: I0202 15:20:35.950966 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2c0443cb-4d5e-4389-a935-1821f94f97a3-serving-cert\") pod \"controller-manager-86cb48757f-sfmxz\" (UID: \"2c0443cb-4d5e-4389-a935-1821f94f97a3\") " pod="openshift-controller-manager/controller-manager-86cb48757f-sfmxz" Feb 02 15:20:35 crc kubenswrapper[4733]: I0202 15:20:35.951012 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7718b7b0-f1a7-4b87-8ebe-345e9100463d-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7718b7b0-f1a7-4b87-8ebe-345e9100463d" (UID: "7718b7b0-f1a7-4b87-8ebe-345e9100463d"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:20:35 crc kubenswrapper[4733]: I0202 15:20:35.951054 4733 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7718b7b0-f1a7-4b87-8ebe-345e9100463d-client-ca\") on node \"crc\" DevicePath \"\"" Feb 02 15:20:35 crc kubenswrapper[4733]: I0202 15:20:35.951071 4733 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f7829943-cabc-49c5-a547-62124e38d49e-client-ca\") on node \"crc\" DevicePath \"\"" Feb 02 15:20:35 crc kubenswrapper[4733]: I0202 15:20:35.951084 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7829943-cabc-49c5-a547-62124e38d49e-config\") on node \"crc\" DevicePath \"\"" Feb 02 15:20:35 crc kubenswrapper[4733]: I0202 15:20:35.951195 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7718b7b0-f1a7-4b87-8ebe-345e9100463d-config" (OuterVolumeSpecName: "config") pod "7718b7b0-f1a7-4b87-8ebe-345e9100463d" (UID: "7718b7b0-f1a7-4b87-8ebe-345e9100463d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:20:35 crc kubenswrapper[4733]: I0202 15:20:35.955487 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7718b7b0-f1a7-4b87-8ebe-345e9100463d-kube-api-access-vfvl2" (OuterVolumeSpecName: "kube-api-access-vfvl2") pod "7718b7b0-f1a7-4b87-8ebe-345e9100463d" (UID: "7718b7b0-f1a7-4b87-8ebe-345e9100463d"). InnerVolumeSpecName "kube-api-access-vfvl2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:20:35 crc kubenswrapper[4733]: I0202 15:20:35.955719 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7718b7b0-f1a7-4b87-8ebe-345e9100463d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7718b7b0-f1a7-4b87-8ebe-345e9100463d" (UID: "7718b7b0-f1a7-4b87-8ebe-345e9100463d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:20:35 crc kubenswrapper[4733]: I0202 15:20:35.957405 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7829943-cabc-49c5-a547-62124e38d49e-kube-api-access-5f8gt" (OuterVolumeSpecName: "kube-api-access-5f8gt") pod "f7829943-cabc-49c5-a547-62124e38d49e" (UID: "f7829943-cabc-49c5-a547-62124e38d49e"). InnerVolumeSpecName "kube-api-access-5f8gt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:20:35 crc kubenswrapper[4733]: I0202 15:20:35.957444 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7829943-cabc-49c5-a547-62124e38d49e-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "f7829943-cabc-49c5-a547-62124e38d49e" (UID: "f7829943-cabc-49c5-a547-62124e38d49e"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:20:36 crc kubenswrapper[4733]: I0202 15:20:36.052475 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c0443cb-4d5e-4389-a935-1821f94f97a3-config\") pod \"controller-manager-86cb48757f-sfmxz\" (UID: \"2c0443cb-4d5e-4389-a935-1821f94f97a3\") " pod="openshift-controller-manager/controller-manager-86cb48757f-sfmxz" Feb 02 15:20:36 crc kubenswrapper[4733]: I0202 15:20:36.052559 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2c0443cb-4d5e-4389-a935-1821f94f97a3-serving-cert\") pod \"controller-manager-86cb48757f-sfmxz\" (UID: \"2c0443cb-4d5e-4389-a935-1821f94f97a3\") " pod="openshift-controller-manager/controller-manager-86cb48757f-sfmxz" Feb 02 15:20:36 crc kubenswrapper[4733]: I0202 15:20:36.052614 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2c0443cb-4d5e-4389-a935-1821f94f97a3-client-ca\") pod \"controller-manager-86cb48757f-sfmxz\" (UID: \"2c0443cb-4d5e-4389-a935-1821f94f97a3\") " pod="openshift-controller-manager/controller-manager-86cb48757f-sfmxz" Feb 02 15:20:36 crc kubenswrapper[4733]: I0202 15:20:36.052662 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2p6h\" (UniqueName: \"kubernetes.io/projected/2c0443cb-4d5e-4389-a935-1821f94f97a3-kube-api-access-r2p6h\") pod \"controller-manager-86cb48757f-sfmxz\" (UID: \"2c0443cb-4d5e-4389-a935-1821f94f97a3\") " pod="openshift-controller-manager/controller-manager-86cb48757f-sfmxz" Feb 02 15:20:36 crc kubenswrapper[4733]: I0202 15:20:36.053311 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2c0443cb-4d5e-4389-a935-1821f94f97a3-proxy-ca-bundles\") pod \"controller-manager-86cb48757f-sfmxz\" (UID: \"2c0443cb-4d5e-4389-a935-1821f94f97a3\") " pod="openshift-controller-manager/controller-manager-86cb48757f-sfmxz" Feb 02 15:20:36 crc kubenswrapper[4733]: I0202 15:20:36.053462 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7829943-cabc-49c5-a547-62124e38d49e-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 15:20:36 crc kubenswrapper[4733]: I0202 15:20:36.053501 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7718b7b0-f1a7-4b87-8ebe-345e9100463d-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 15:20:36 crc kubenswrapper[4733]: I0202 15:20:36.053524 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5f8gt\" (UniqueName: \"kubernetes.io/projected/f7829943-cabc-49c5-a547-62124e38d49e-kube-api-access-5f8gt\") on node \"crc\" DevicePath \"\"" Feb 02 15:20:36 crc kubenswrapper[4733]: I0202 15:20:36.053545 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7718b7b0-f1a7-4b87-8ebe-345e9100463d-config\") on node \"crc\" DevicePath \"\"" Feb 02 15:20:36 crc kubenswrapper[4733]: I0202 15:20:36.053617 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vfvl2\" (UniqueName: \"kubernetes.io/projected/7718b7b0-f1a7-4b87-8ebe-345e9100463d-kube-api-access-vfvl2\") on node \"crc\" DevicePath \"\"" Feb 02 15:20:36 crc kubenswrapper[4733]: I0202 15:20:36.053636 4733 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7718b7b0-f1a7-4b87-8ebe-345e9100463d-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 02 15:20:36 crc kubenswrapper[4733]: I0202 15:20:36.054303 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2c0443cb-4d5e-4389-a935-1821f94f97a3-client-ca\") pod \"controller-manager-86cb48757f-sfmxz\" (UID: \"2c0443cb-4d5e-4389-a935-1821f94f97a3\") " pod="openshift-controller-manager/controller-manager-86cb48757f-sfmxz" Feb 02 15:20:36 crc kubenswrapper[4733]: I0202 15:20:36.054959 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c0443cb-4d5e-4389-a935-1821f94f97a3-config\") pod \"controller-manager-86cb48757f-sfmxz\" (UID: \"2c0443cb-4d5e-4389-a935-1821f94f97a3\") " pod="openshift-controller-manager/controller-manager-86cb48757f-sfmxz" Feb 02 15:20:36 crc kubenswrapper[4733]: I0202 15:20:36.055753 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2c0443cb-4d5e-4389-a935-1821f94f97a3-proxy-ca-bundles\") pod \"controller-manager-86cb48757f-sfmxz\" (UID: \"2c0443cb-4d5e-4389-a935-1821f94f97a3\") " pod="openshift-controller-manager/controller-manager-86cb48757f-sfmxz" Feb 02 15:20:36 crc kubenswrapper[4733]: I0202 15:20:36.056997 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2c0443cb-4d5e-4389-a935-1821f94f97a3-serving-cert\") pod \"controller-manager-86cb48757f-sfmxz\" (UID: \"2c0443cb-4d5e-4389-a935-1821f94f97a3\") " pod="openshift-controller-manager/controller-manager-86cb48757f-sfmxz" Feb 02 15:20:36 crc kubenswrapper[4733]: I0202 15:20:36.071376 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2p6h\" (UniqueName: \"kubernetes.io/projected/2c0443cb-4d5e-4389-a935-1821f94f97a3-kube-api-access-r2p6h\") pod \"controller-manager-86cb48757f-sfmxz\" (UID: \"2c0443cb-4d5e-4389-a935-1821f94f97a3\") " pod="openshift-controller-manager/controller-manager-86cb48757f-sfmxz" Feb 02 15:20:36 crc kubenswrapper[4733]: I0202 15:20:36.183155 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-86cb48757f-sfmxz" Feb 02 15:20:36 crc kubenswrapper[4733]: I0202 15:20:36.462326 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-86cb48757f-sfmxz"] Feb 02 15:20:36 crc kubenswrapper[4733]: I0202 15:20:36.488277 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-86cb48757f-sfmxz" event={"ID":"2c0443cb-4d5e-4389-a935-1821f94f97a3","Type":"ContainerStarted","Data":"43a4946fe349d684ad1fa8f79ab1903d454de1edeb168690d5d27bb8f3df8b15"} Feb 02 15:20:36 crc kubenswrapper[4733]: I0202 15:20:36.489632 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-98h6l" event={"ID":"f7829943-cabc-49c5-a547-62124e38d49e","Type":"ContainerDied","Data":"69b3eabefd5f9ed969de63b29afef1d1720ae7a64d3b673e908abe27a6e2fc02"} Feb 02 15:20:36 crc kubenswrapper[4733]: I0202 15:20:36.489657 4733 scope.go:117] "RemoveContainer" containerID="918ae4dd606f38b41b783f31ac022be12bb0fe2af0666b1799e8dcec1e790b74" Feb 02 15:20:36 crc kubenswrapper[4733]: I0202 15:20:36.489728 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-98h6l" Feb 02 15:20:36 crc kubenswrapper[4733]: I0202 15:20:36.494074 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-m62b2" event={"ID":"7718b7b0-f1a7-4b87-8ebe-345e9100463d","Type":"ContainerDied","Data":"52c3bfd7078308d02908ed10f8d54d9f00bd3465c4fc389e2c149c066cbec839"} Feb 02 15:20:36 crc kubenswrapper[4733]: I0202 15:20:36.494154 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-m62b2" Feb 02 15:20:36 crc kubenswrapper[4733]: I0202 15:20:36.521127 4733 scope.go:117] "RemoveContainer" containerID="e1d6f6e0975a0529556931f0c74afc5d70e93d838245956f2238a399298a6400" Feb 02 15:20:36 crc kubenswrapper[4733]: I0202 15:20:36.522195 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-98h6l"] Feb 02 15:20:36 crc kubenswrapper[4733]: I0202 15:20:36.527362 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-98h6l"] Feb 02 15:20:36 crc kubenswrapper[4733]: I0202 15:20:36.541748 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-m62b2"] Feb 02 15:20:36 crc kubenswrapper[4733]: I0202 15:20:36.547470 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-m62b2"] Feb 02 15:20:37 crc kubenswrapper[4733]: I0202 15:20:37.271923 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7718b7b0-f1a7-4b87-8ebe-345e9100463d" path="/var/lib/kubelet/pods/7718b7b0-f1a7-4b87-8ebe-345e9100463d/volumes" Feb 02 15:20:37 crc kubenswrapper[4733]: I0202 15:20:37.272475 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7829943-cabc-49c5-a547-62124e38d49e" path="/var/lib/kubelet/pods/f7829943-cabc-49c5-a547-62124e38d49e/volumes" Feb 02 15:20:37 crc kubenswrapper[4733]: I0202 15:20:37.499583 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-86cb48757f-sfmxz" event={"ID":"2c0443cb-4d5e-4389-a935-1821f94f97a3","Type":"ContainerStarted","Data":"a778bb87432a7764f6fdf7d57a75918425aa1dbafc7e3e60937f35dfe56b8529"} Feb 02 15:20:37 crc kubenswrapper[4733]: I0202 15:20:37.500553 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-86cb48757f-sfmxz" Feb 02 15:20:37 crc kubenswrapper[4733]: I0202 15:20:37.505785 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-86cb48757f-sfmxz" Feb 02 15:20:37 crc kubenswrapper[4733]: I0202 15:20:37.517019 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-86cb48757f-sfmxz" podStartSLOduration=5.5170053679999995 podStartE2EDuration="5.517005368s" podCreationTimestamp="2026-02-02 15:20:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:20:37.513339308 +0000 UTC m=+320.964800676" watchObservedRunningTime="2026-02-02 15:20:37.517005368 +0000 UTC m=+320.968466736" Feb 02 15:20:38 crc kubenswrapper[4733]: I0202 15:20:38.046188 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-86cb48757f-sfmxz"] Feb 02 15:20:38 crc kubenswrapper[4733]: I0202 15:20:38.084131 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-747d45d6f5-stjql"] Feb 02 15:20:38 crc kubenswrapper[4733]: I0202 15:20:38.085008 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-747d45d6f5-stjql" Feb 02 15:20:38 crc kubenswrapper[4733]: I0202 15:20:38.087374 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 02 15:20:38 crc kubenswrapper[4733]: I0202 15:20:38.087441 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 02 15:20:38 crc kubenswrapper[4733]: I0202 15:20:38.087717 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 02 15:20:38 crc kubenswrapper[4733]: I0202 15:20:38.087887 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 02 15:20:38 crc kubenswrapper[4733]: I0202 15:20:38.088294 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 02 15:20:38 crc kubenswrapper[4733]: I0202 15:20:38.091556 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 02 15:20:38 crc kubenswrapper[4733]: I0202 15:20:38.096317 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-747d45d6f5-stjql"] Feb 02 15:20:38 crc kubenswrapper[4733]: I0202 15:20:38.183184 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5d8a74eb-c898-4077-b57b-72eddb0c958c-serving-cert\") pod \"route-controller-manager-747d45d6f5-stjql\" (UID: \"5d8a74eb-c898-4077-b57b-72eddb0c958c\") " pod="openshift-route-controller-manager/route-controller-manager-747d45d6f5-stjql" Feb 02 15:20:38 crc kubenswrapper[4733]: I0202 15:20:38.183247 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d8a74eb-c898-4077-b57b-72eddb0c958c-config\") pod \"route-controller-manager-747d45d6f5-stjql\" (UID: \"5d8a74eb-c898-4077-b57b-72eddb0c958c\") " pod="openshift-route-controller-manager/route-controller-manager-747d45d6f5-stjql" Feb 02 15:20:38 crc kubenswrapper[4733]: I0202 15:20:38.183280 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5d8a74eb-c898-4077-b57b-72eddb0c958c-client-ca\") pod \"route-controller-manager-747d45d6f5-stjql\" (UID: \"5d8a74eb-c898-4077-b57b-72eddb0c958c\") " pod="openshift-route-controller-manager/route-controller-manager-747d45d6f5-stjql" Feb 02 15:20:38 crc kubenswrapper[4733]: I0202 15:20:38.183512 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lts4f\" (UniqueName: \"kubernetes.io/projected/5d8a74eb-c898-4077-b57b-72eddb0c958c-kube-api-access-lts4f\") pod \"route-controller-manager-747d45d6f5-stjql\" (UID: \"5d8a74eb-c898-4077-b57b-72eddb0c958c\") " pod="openshift-route-controller-manager/route-controller-manager-747d45d6f5-stjql" Feb 02 15:20:38 crc kubenswrapper[4733]: I0202 15:20:38.285626 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lts4f\" (UniqueName: \"kubernetes.io/projected/5d8a74eb-c898-4077-b57b-72eddb0c958c-kube-api-access-lts4f\") pod \"route-controller-manager-747d45d6f5-stjql\" (UID: \"5d8a74eb-c898-4077-b57b-72eddb0c958c\") " pod="openshift-route-controller-manager/route-controller-manager-747d45d6f5-stjql" Feb 02 15:20:38 crc kubenswrapper[4733]: I0202 15:20:38.285773 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5d8a74eb-c898-4077-b57b-72eddb0c958c-serving-cert\") pod \"route-controller-manager-747d45d6f5-stjql\" (UID: \"5d8a74eb-c898-4077-b57b-72eddb0c958c\") " pod="openshift-route-controller-manager/route-controller-manager-747d45d6f5-stjql" Feb 02 15:20:38 crc kubenswrapper[4733]: I0202 15:20:38.285885 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d8a74eb-c898-4077-b57b-72eddb0c958c-config\") pod \"route-controller-manager-747d45d6f5-stjql\" (UID: \"5d8a74eb-c898-4077-b57b-72eddb0c958c\") " pod="openshift-route-controller-manager/route-controller-manager-747d45d6f5-stjql" Feb 02 15:20:38 crc kubenswrapper[4733]: I0202 15:20:38.285955 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5d8a74eb-c898-4077-b57b-72eddb0c958c-client-ca\") pod \"route-controller-manager-747d45d6f5-stjql\" (UID: \"5d8a74eb-c898-4077-b57b-72eddb0c958c\") " pod="openshift-route-controller-manager/route-controller-manager-747d45d6f5-stjql" Feb 02 15:20:38 crc kubenswrapper[4733]: I0202 15:20:38.287124 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d8a74eb-c898-4077-b57b-72eddb0c958c-config\") pod \"route-controller-manager-747d45d6f5-stjql\" (UID: \"5d8a74eb-c898-4077-b57b-72eddb0c958c\") " pod="openshift-route-controller-manager/route-controller-manager-747d45d6f5-stjql" Feb 02 15:20:38 crc kubenswrapper[4733]: I0202 15:20:38.299289 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5d8a74eb-c898-4077-b57b-72eddb0c958c-client-ca\") pod \"route-controller-manager-747d45d6f5-stjql\" (UID: \"5d8a74eb-c898-4077-b57b-72eddb0c958c\") " pod="openshift-route-controller-manager/route-controller-manager-747d45d6f5-stjql" Feb 02 15:20:38 crc kubenswrapper[4733]: I0202 15:20:38.299798 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5d8a74eb-c898-4077-b57b-72eddb0c958c-serving-cert\") pod \"route-controller-manager-747d45d6f5-stjql\" (UID: \"5d8a74eb-c898-4077-b57b-72eddb0c958c\") " pod="openshift-route-controller-manager/route-controller-manager-747d45d6f5-stjql" Feb 02 15:20:38 crc kubenswrapper[4733]: I0202 15:20:38.309060 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-747d45d6f5-stjql"] Feb 02 15:20:38 crc kubenswrapper[4733]: E0202 15:20:38.309694 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-lts4f], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openshift-route-controller-manager/route-controller-manager-747d45d6f5-stjql" podUID="5d8a74eb-c898-4077-b57b-72eddb0c958c" Feb 02 15:20:38 crc kubenswrapper[4733]: I0202 15:20:38.311214 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lts4f\" (UniqueName: \"kubernetes.io/projected/5d8a74eb-c898-4077-b57b-72eddb0c958c-kube-api-access-lts4f\") pod \"route-controller-manager-747d45d6f5-stjql\" (UID: \"5d8a74eb-c898-4077-b57b-72eddb0c958c\") " pod="openshift-route-controller-manager/route-controller-manager-747d45d6f5-stjql" Feb 02 15:20:38 crc kubenswrapper[4733]: I0202 15:20:38.507089 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-747d45d6f5-stjql" Feb 02 15:20:38 crc kubenswrapper[4733]: I0202 15:20:38.514961 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-747d45d6f5-stjql" Feb 02 15:20:38 crc kubenswrapper[4733]: I0202 15:20:38.589030 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5d8a74eb-c898-4077-b57b-72eddb0c958c-client-ca\") pod \"5d8a74eb-c898-4077-b57b-72eddb0c958c\" (UID: \"5d8a74eb-c898-4077-b57b-72eddb0c958c\") " Feb 02 15:20:38 crc kubenswrapper[4733]: I0202 15:20:38.589108 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5d8a74eb-c898-4077-b57b-72eddb0c958c-serving-cert\") pod \"5d8a74eb-c898-4077-b57b-72eddb0c958c\" (UID: \"5d8a74eb-c898-4077-b57b-72eddb0c958c\") " Feb 02 15:20:38 crc kubenswrapper[4733]: I0202 15:20:38.589269 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lts4f\" (UniqueName: \"kubernetes.io/projected/5d8a74eb-c898-4077-b57b-72eddb0c958c-kube-api-access-lts4f\") pod \"5d8a74eb-c898-4077-b57b-72eddb0c958c\" (UID: \"5d8a74eb-c898-4077-b57b-72eddb0c958c\") " Feb 02 15:20:38 crc kubenswrapper[4733]: I0202 15:20:38.589354 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d8a74eb-c898-4077-b57b-72eddb0c958c-config\") pod \"5d8a74eb-c898-4077-b57b-72eddb0c958c\" (UID: \"5d8a74eb-c898-4077-b57b-72eddb0c958c\") " Feb 02 15:20:38 crc kubenswrapper[4733]: I0202 15:20:38.589422 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d8a74eb-c898-4077-b57b-72eddb0c958c-client-ca" (OuterVolumeSpecName: "client-ca") pod "5d8a74eb-c898-4077-b57b-72eddb0c958c" (UID: "5d8a74eb-c898-4077-b57b-72eddb0c958c"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:20:38 crc kubenswrapper[4733]: I0202 15:20:38.589790 4733 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5d8a74eb-c898-4077-b57b-72eddb0c958c-client-ca\") on node \"crc\" DevicePath \"\"" Feb 02 15:20:38 crc kubenswrapper[4733]: I0202 15:20:38.590975 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d8a74eb-c898-4077-b57b-72eddb0c958c-config" (OuterVolumeSpecName: "config") pod "5d8a74eb-c898-4077-b57b-72eddb0c958c" (UID: "5d8a74eb-c898-4077-b57b-72eddb0c958c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:20:38 crc kubenswrapper[4733]: I0202 15:20:38.592412 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d8a74eb-c898-4077-b57b-72eddb0c958c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5d8a74eb-c898-4077-b57b-72eddb0c958c" (UID: "5d8a74eb-c898-4077-b57b-72eddb0c958c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:20:38 crc kubenswrapper[4733]: I0202 15:20:38.594668 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d8a74eb-c898-4077-b57b-72eddb0c958c-kube-api-access-lts4f" (OuterVolumeSpecName: "kube-api-access-lts4f") pod "5d8a74eb-c898-4077-b57b-72eddb0c958c" (UID: "5d8a74eb-c898-4077-b57b-72eddb0c958c"). InnerVolumeSpecName "kube-api-access-lts4f". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:20:38 crc kubenswrapper[4733]: I0202 15:20:38.690713 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d8a74eb-c898-4077-b57b-72eddb0c958c-config\") on node \"crc\" DevicePath \"\"" Feb 02 15:20:38 crc kubenswrapper[4733]: I0202 15:20:38.690754 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5d8a74eb-c898-4077-b57b-72eddb0c958c-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 15:20:38 crc kubenswrapper[4733]: I0202 15:20:38.690768 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lts4f\" (UniqueName: \"kubernetes.io/projected/5d8a74eb-c898-4077-b57b-72eddb0c958c-kube-api-access-lts4f\") on node \"crc\" DevicePath \"\"" Feb 02 15:20:39 crc kubenswrapper[4733]: I0202 15:20:39.512572 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-747d45d6f5-stjql" Feb 02 15:20:39 crc kubenswrapper[4733]: I0202 15:20:39.512666 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-86cb48757f-sfmxz" podUID="2c0443cb-4d5e-4389-a935-1821f94f97a3" containerName="controller-manager" containerID="cri-o://a778bb87432a7764f6fdf7d57a75918425aa1dbafc7e3e60937f35dfe56b8529" gracePeriod=30 Feb 02 15:20:39 crc kubenswrapper[4733]: I0202 15:20:39.552513 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-747d45d6f5-stjql"] Feb 02 15:20:39 crc kubenswrapper[4733]: I0202 15:20:39.556298 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5cf5674648-xw84g"] Feb 02 15:20:39 crc kubenswrapper[4733]: I0202 15:20:39.557780 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5cf5674648-xw84g" Feb 02 15:20:39 crc kubenswrapper[4733]: I0202 15:20:39.560678 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-747d45d6f5-stjql"] Feb 02 15:20:39 crc kubenswrapper[4733]: I0202 15:20:39.560703 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 02 15:20:39 crc kubenswrapper[4733]: I0202 15:20:39.561107 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 02 15:20:39 crc kubenswrapper[4733]: I0202 15:20:39.561281 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 02 15:20:39 crc kubenswrapper[4733]: I0202 15:20:39.561308 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 02 15:20:39 crc kubenswrapper[4733]: I0202 15:20:39.561446 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 02 15:20:39 crc kubenswrapper[4733]: I0202 15:20:39.563480 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 02 15:20:39 crc kubenswrapper[4733]: I0202 15:20:39.604282 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ba456f3a-c060-4c07-b370-521e08c4f57a-serving-cert\") pod \"route-controller-manager-5cf5674648-xw84g\" (UID: \"ba456f3a-c060-4c07-b370-521e08c4f57a\") " pod="openshift-route-controller-manager/route-controller-manager-5cf5674648-xw84g" Feb 02 15:20:39 crc kubenswrapper[4733]: I0202 15:20:39.604398 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-574b2\" (UniqueName: \"kubernetes.io/projected/ba456f3a-c060-4c07-b370-521e08c4f57a-kube-api-access-574b2\") pod \"route-controller-manager-5cf5674648-xw84g\" (UID: \"ba456f3a-c060-4c07-b370-521e08c4f57a\") " pod="openshift-route-controller-manager/route-controller-manager-5cf5674648-xw84g" Feb 02 15:20:39 crc kubenswrapper[4733]: I0202 15:20:39.604480 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba456f3a-c060-4c07-b370-521e08c4f57a-config\") pod \"route-controller-manager-5cf5674648-xw84g\" (UID: \"ba456f3a-c060-4c07-b370-521e08c4f57a\") " pod="openshift-route-controller-manager/route-controller-manager-5cf5674648-xw84g" Feb 02 15:20:39 crc kubenswrapper[4733]: I0202 15:20:39.604534 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ba456f3a-c060-4c07-b370-521e08c4f57a-client-ca\") pod \"route-controller-manager-5cf5674648-xw84g\" (UID: \"ba456f3a-c060-4c07-b370-521e08c4f57a\") " pod="openshift-route-controller-manager/route-controller-manager-5cf5674648-xw84g" Feb 02 15:20:39 crc kubenswrapper[4733]: I0202 15:20:39.609570 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5cf5674648-xw84g"] Feb 02 15:20:39 crc kubenswrapper[4733]: I0202 15:20:39.706406 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ba456f3a-c060-4c07-b370-521e08c4f57a-serving-cert\") pod \"route-controller-manager-5cf5674648-xw84g\" (UID: \"ba456f3a-c060-4c07-b370-521e08c4f57a\") " pod="openshift-route-controller-manager/route-controller-manager-5cf5674648-xw84g" Feb 02 15:20:39 crc kubenswrapper[4733]: I0202 15:20:39.706496 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-574b2\" (UniqueName: \"kubernetes.io/projected/ba456f3a-c060-4c07-b370-521e08c4f57a-kube-api-access-574b2\") pod \"route-controller-manager-5cf5674648-xw84g\" (UID: \"ba456f3a-c060-4c07-b370-521e08c4f57a\") " pod="openshift-route-controller-manager/route-controller-manager-5cf5674648-xw84g" Feb 02 15:20:39 crc kubenswrapper[4733]: I0202 15:20:39.706551 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba456f3a-c060-4c07-b370-521e08c4f57a-config\") pod \"route-controller-manager-5cf5674648-xw84g\" (UID: \"ba456f3a-c060-4c07-b370-521e08c4f57a\") " pod="openshift-route-controller-manager/route-controller-manager-5cf5674648-xw84g" Feb 02 15:20:39 crc kubenswrapper[4733]: I0202 15:20:39.706585 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ba456f3a-c060-4c07-b370-521e08c4f57a-client-ca\") pod \"route-controller-manager-5cf5674648-xw84g\" (UID: \"ba456f3a-c060-4c07-b370-521e08c4f57a\") " pod="openshift-route-controller-manager/route-controller-manager-5cf5674648-xw84g" Feb 02 15:20:39 crc kubenswrapper[4733]: I0202 15:20:39.708058 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ba456f3a-c060-4c07-b370-521e08c4f57a-client-ca\") pod \"route-controller-manager-5cf5674648-xw84g\" (UID: \"ba456f3a-c060-4c07-b370-521e08c4f57a\") " pod="openshift-route-controller-manager/route-controller-manager-5cf5674648-xw84g" Feb 02 15:20:39 crc kubenswrapper[4733]: I0202 15:20:39.708681 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba456f3a-c060-4c07-b370-521e08c4f57a-config\") pod \"route-controller-manager-5cf5674648-xw84g\" (UID: \"ba456f3a-c060-4c07-b370-521e08c4f57a\") " pod="openshift-route-controller-manager/route-controller-manager-5cf5674648-xw84g" Feb 02 15:20:39 crc kubenswrapper[4733]: I0202 15:20:39.712563 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ba456f3a-c060-4c07-b370-521e08c4f57a-serving-cert\") pod \"route-controller-manager-5cf5674648-xw84g\" (UID: \"ba456f3a-c060-4c07-b370-521e08c4f57a\") " pod="openshift-route-controller-manager/route-controller-manager-5cf5674648-xw84g" Feb 02 15:20:39 crc kubenswrapper[4733]: I0202 15:20:39.723121 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-574b2\" (UniqueName: \"kubernetes.io/projected/ba456f3a-c060-4c07-b370-521e08c4f57a-kube-api-access-574b2\") pod \"route-controller-manager-5cf5674648-xw84g\" (UID: \"ba456f3a-c060-4c07-b370-521e08c4f57a\") " pod="openshift-route-controller-manager/route-controller-manager-5cf5674648-xw84g" Feb 02 15:20:39 crc kubenswrapper[4733]: I0202 15:20:39.887837 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5cf5674648-xw84g" Feb 02 15:20:40 crc kubenswrapper[4733]: I0202 15:20:40.784775 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5cf5674648-xw84g"] Feb 02 15:20:40 crc kubenswrapper[4733]: W0202 15:20:40.799694 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podba456f3a_c060_4c07_b370_521e08c4f57a.slice/crio-c581c02249b43aed6bbfdb1458f38591321224b89ff14be0ab7a4f41795960b5 WatchSource:0}: Error finding container c581c02249b43aed6bbfdb1458f38591321224b89ff14be0ab7a4f41795960b5: Status 404 returned error can't find the container with id c581c02249b43aed6bbfdb1458f38591321224b89ff14be0ab7a4f41795960b5 Feb 02 15:20:41 crc kubenswrapper[4733]: I0202 15:20:41.268101 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d8a74eb-c898-4077-b57b-72eddb0c958c" path="/var/lib/kubelet/pods/5d8a74eb-c898-4077-b57b-72eddb0c958c/volumes" Feb 02 15:20:41 crc kubenswrapper[4733]: I0202 15:20:41.526280 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5cf5674648-xw84g" event={"ID":"ba456f3a-c060-4c07-b370-521e08c4f57a","Type":"ContainerStarted","Data":"4abfe5d8edf004c3326af0a64d1160b23ef5ce149e733ee50e3aa121fb1585f2"} Feb 02 15:20:41 crc kubenswrapper[4733]: I0202 15:20:41.526618 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5cf5674648-xw84g" event={"ID":"ba456f3a-c060-4c07-b370-521e08c4f57a","Type":"ContainerStarted","Data":"c581c02249b43aed6bbfdb1458f38591321224b89ff14be0ab7a4f41795960b5"} Feb 02 15:20:41 crc kubenswrapper[4733]: I0202 15:20:41.528578 4733 generic.go:334] "Generic (PLEG): container finished" podID="2c0443cb-4d5e-4389-a935-1821f94f97a3" containerID="a778bb87432a7764f6fdf7d57a75918425aa1dbafc7e3e60937f35dfe56b8529" exitCode=0 Feb 02 15:20:41 crc kubenswrapper[4733]: I0202 15:20:41.528625 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-86cb48757f-sfmxz" event={"ID":"2c0443cb-4d5e-4389-a935-1821f94f97a3","Type":"ContainerDied","Data":"a778bb87432a7764f6fdf7d57a75918425aa1dbafc7e3e60937f35dfe56b8529"} Feb 02 15:20:42 crc kubenswrapper[4733]: I0202 15:20:42.323642 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-86cb48757f-sfmxz" Feb 02 15:20:42 crc kubenswrapper[4733]: I0202 15:20:42.344492 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2c0443cb-4d5e-4389-a935-1821f94f97a3-client-ca\") pod \"2c0443cb-4d5e-4389-a935-1821f94f97a3\" (UID: \"2c0443cb-4d5e-4389-a935-1821f94f97a3\") " Feb 02 15:20:42 crc kubenswrapper[4733]: I0202 15:20:42.344572 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r2p6h\" (UniqueName: \"kubernetes.io/projected/2c0443cb-4d5e-4389-a935-1821f94f97a3-kube-api-access-r2p6h\") pod \"2c0443cb-4d5e-4389-a935-1821f94f97a3\" (UID: \"2c0443cb-4d5e-4389-a935-1821f94f97a3\") " Feb 02 15:20:42 crc kubenswrapper[4733]: I0202 15:20:42.344600 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c0443cb-4d5e-4389-a935-1821f94f97a3-config\") pod \"2c0443cb-4d5e-4389-a935-1821f94f97a3\" (UID: \"2c0443cb-4d5e-4389-a935-1821f94f97a3\") " Feb 02 15:20:42 crc kubenswrapper[4733]: I0202 15:20:42.344622 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2c0443cb-4d5e-4389-a935-1821f94f97a3-serving-cert\") pod \"2c0443cb-4d5e-4389-a935-1821f94f97a3\" (UID: \"2c0443cb-4d5e-4389-a935-1821f94f97a3\") " Feb 02 15:20:42 crc kubenswrapper[4733]: I0202 15:20:42.344650 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2c0443cb-4d5e-4389-a935-1821f94f97a3-proxy-ca-bundles\") pod \"2c0443cb-4d5e-4389-a935-1821f94f97a3\" (UID: \"2c0443cb-4d5e-4389-a935-1821f94f97a3\") " Feb 02 15:20:42 crc kubenswrapper[4733]: I0202 15:20:42.345484 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c0443cb-4d5e-4389-a935-1821f94f97a3-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "2c0443cb-4d5e-4389-a935-1821f94f97a3" (UID: "2c0443cb-4d5e-4389-a935-1821f94f97a3"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:20:42 crc kubenswrapper[4733]: I0202 15:20:42.345902 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c0443cb-4d5e-4389-a935-1821f94f97a3-client-ca" (OuterVolumeSpecName: "client-ca") pod "2c0443cb-4d5e-4389-a935-1821f94f97a3" (UID: "2c0443cb-4d5e-4389-a935-1821f94f97a3"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:20:42 crc kubenswrapper[4733]: I0202 15:20:42.346197 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c0443cb-4d5e-4389-a935-1821f94f97a3-config" (OuterVolumeSpecName: "config") pod "2c0443cb-4d5e-4389-a935-1821f94f97a3" (UID: "2c0443cb-4d5e-4389-a935-1821f94f97a3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:20:42 crc kubenswrapper[4733]: I0202 15:20:42.353569 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c0443cb-4d5e-4389-a935-1821f94f97a3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "2c0443cb-4d5e-4389-a935-1821f94f97a3" (UID: "2c0443cb-4d5e-4389-a935-1821f94f97a3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:20:42 crc kubenswrapper[4733]: I0202 15:20:42.354642 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c0443cb-4d5e-4389-a935-1821f94f97a3-kube-api-access-r2p6h" (OuterVolumeSpecName: "kube-api-access-r2p6h") pod "2c0443cb-4d5e-4389-a935-1821f94f97a3" (UID: "2c0443cb-4d5e-4389-a935-1821f94f97a3"). InnerVolumeSpecName "kube-api-access-r2p6h". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:20:42 crc kubenswrapper[4733]: I0202 15:20:42.371933 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-54464559b6-tpnl9"] Feb 02 15:20:42 crc kubenswrapper[4733]: E0202 15:20:42.372272 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c0443cb-4d5e-4389-a935-1821f94f97a3" containerName="controller-manager" Feb 02 15:20:42 crc kubenswrapper[4733]: I0202 15:20:42.372307 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c0443cb-4d5e-4389-a935-1821f94f97a3" containerName="controller-manager" Feb 02 15:20:42 crc kubenswrapper[4733]: I0202 15:20:42.372452 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c0443cb-4d5e-4389-a935-1821f94f97a3" containerName="controller-manager" Feb 02 15:20:42 crc kubenswrapper[4733]: I0202 15:20:42.372978 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-54464559b6-tpnl9" Feb 02 15:20:42 crc kubenswrapper[4733]: I0202 15:20:42.380011 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-54464559b6-tpnl9"] Feb 02 15:20:42 crc kubenswrapper[4733]: I0202 15:20:42.446144 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nj2s4\" (UniqueName: \"kubernetes.io/projected/e1407e69-2431-4fdf-b8a7-b6396b46b6d9-kube-api-access-nj2s4\") pod \"controller-manager-54464559b6-tpnl9\" (UID: \"e1407e69-2431-4fdf-b8a7-b6396b46b6d9\") " pod="openshift-controller-manager/controller-manager-54464559b6-tpnl9" Feb 02 15:20:42 crc kubenswrapper[4733]: I0202 15:20:42.446252 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1407e69-2431-4fdf-b8a7-b6396b46b6d9-config\") pod \"controller-manager-54464559b6-tpnl9\" (UID: \"e1407e69-2431-4fdf-b8a7-b6396b46b6d9\") " pod="openshift-controller-manager/controller-manager-54464559b6-tpnl9" Feb 02 15:20:42 crc kubenswrapper[4733]: I0202 15:20:42.446275 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e1407e69-2431-4fdf-b8a7-b6396b46b6d9-client-ca\") pod \"controller-manager-54464559b6-tpnl9\" (UID: \"e1407e69-2431-4fdf-b8a7-b6396b46b6d9\") " pod="openshift-controller-manager/controller-manager-54464559b6-tpnl9" Feb 02 15:20:42 crc kubenswrapper[4733]: I0202 15:20:42.446302 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e1407e69-2431-4fdf-b8a7-b6396b46b6d9-proxy-ca-bundles\") pod \"controller-manager-54464559b6-tpnl9\" (UID: \"e1407e69-2431-4fdf-b8a7-b6396b46b6d9\") " pod="openshift-controller-manager/controller-manager-54464559b6-tpnl9" Feb 02 15:20:42 crc kubenswrapper[4733]: I0202 15:20:42.446464 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e1407e69-2431-4fdf-b8a7-b6396b46b6d9-serving-cert\") pod \"controller-manager-54464559b6-tpnl9\" (UID: \"e1407e69-2431-4fdf-b8a7-b6396b46b6d9\") " pod="openshift-controller-manager/controller-manager-54464559b6-tpnl9" Feb 02 15:20:42 crc kubenswrapper[4733]: I0202 15:20:42.446601 4733 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2c0443cb-4d5e-4389-a935-1821f94f97a3-client-ca\") on node \"crc\" DevicePath \"\"" Feb 02 15:20:42 crc kubenswrapper[4733]: I0202 15:20:42.446622 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r2p6h\" (UniqueName: \"kubernetes.io/projected/2c0443cb-4d5e-4389-a935-1821f94f97a3-kube-api-access-r2p6h\") on node \"crc\" DevicePath \"\"" Feb 02 15:20:42 crc kubenswrapper[4733]: I0202 15:20:42.446631 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c0443cb-4d5e-4389-a935-1821f94f97a3-config\") on node \"crc\" DevicePath \"\"" Feb 02 15:20:42 crc kubenswrapper[4733]: I0202 15:20:42.446640 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2c0443cb-4d5e-4389-a935-1821f94f97a3-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 15:20:42 crc kubenswrapper[4733]: I0202 15:20:42.446649 4733 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2c0443cb-4d5e-4389-a935-1821f94f97a3-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 02 15:20:42 crc kubenswrapper[4733]: I0202 15:20:42.533863 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-86cb48757f-sfmxz" event={"ID":"2c0443cb-4d5e-4389-a935-1821f94f97a3","Type":"ContainerDied","Data":"43a4946fe349d684ad1fa8f79ab1903d454de1edeb168690d5d27bb8f3df8b15"} Feb 02 15:20:42 crc kubenswrapper[4733]: I0202 15:20:42.533929 4733 scope.go:117] "RemoveContainer" containerID="a778bb87432a7764f6fdf7d57a75918425aa1dbafc7e3e60937f35dfe56b8529" Feb 02 15:20:42 crc kubenswrapper[4733]: I0202 15:20:42.534095 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5cf5674648-xw84g" Feb 02 15:20:42 crc kubenswrapper[4733]: I0202 15:20:42.534285 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-86cb48757f-sfmxz" Feb 02 15:20:42 crc kubenswrapper[4733]: I0202 15:20:42.541680 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5cf5674648-xw84g" Feb 02 15:20:42 crc kubenswrapper[4733]: I0202 15:20:42.548001 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1407e69-2431-4fdf-b8a7-b6396b46b6d9-config\") pod \"controller-manager-54464559b6-tpnl9\" (UID: \"e1407e69-2431-4fdf-b8a7-b6396b46b6d9\") " pod="openshift-controller-manager/controller-manager-54464559b6-tpnl9" Feb 02 15:20:42 crc kubenswrapper[4733]: I0202 15:20:42.548078 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e1407e69-2431-4fdf-b8a7-b6396b46b6d9-client-ca\") pod \"controller-manager-54464559b6-tpnl9\" (UID: \"e1407e69-2431-4fdf-b8a7-b6396b46b6d9\") " pod="openshift-controller-manager/controller-manager-54464559b6-tpnl9" Feb 02 15:20:42 crc kubenswrapper[4733]: I0202 15:20:42.548140 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e1407e69-2431-4fdf-b8a7-b6396b46b6d9-proxy-ca-bundles\") pod \"controller-manager-54464559b6-tpnl9\" (UID: \"e1407e69-2431-4fdf-b8a7-b6396b46b6d9\") " pod="openshift-controller-manager/controller-manager-54464559b6-tpnl9" Feb 02 15:20:42 crc kubenswrapper[4733]: I0202 15:20:42.548266 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e1407e69-2431-4fdf-b8a7-b6396b46b6d9-serving-cert\") pod \"controller-manager-54464559b6-tpnl9\" (UID: \"e1407e69-2431-4fdf-b8a7-b6396b46b6d9\") " pod="openshift-controller-manager/controller-manager-54464559b6-tpnl9" Feb 02 15:20:42 crc kubenswrapper[4733]: I0202 15:20:42.548326 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nj2s4\" (UniqueName: \"kubernetes.io/projected/e1407e69-2431-4fdf-b8a7-b6396b46b6d9-kube-api-access-nj2s4\") pod \"controller-manager-54464559b6-tpnl9\" (UID: \"e1407e69-2431-4fdf-b8a7-b6396b46b6d9\") " pod="openshift-controller-manager/controller-manager-54464559b6-tpnl9" Feb 02 15:20:42 crc kubenswrapper[4733]: I0202 15:20:42.548976 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e1407e69-2431-4fdf-b8a7-b6396b46b6d9-client-ca\") pod \"controller-manager-54464559b6-tpnl9\" (UID: \"e1407e69-2431-4fdf-b8a7-b6396b46b6d9\") " pod="openshift-controller-manager/controller-manager-54464559b6-tpnl9" Feb 02 15:20:42 crc kubenswrapper[4733]: I0202 15:20:42.549199 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1407e69-2431-4fdf-b8a7-b6396b46b6d9-config\") pod \"controller-manager-54464559b6-tpnl9\" (UID: \"e1407e69-2431-4fdf-b8a7-b6396b46b6d9\") " pod="openshift-controller-manager/controller-manager-54464559b6-tpnl9" Feb 02 15:20:42 crc kubenswrapper[4733]: I0202 15:20:42.551806 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e1407e69-2431-4fdf-b8a7-b6396b46b6d9-proxy-ca-bundles\") pod \"controller-manager-54464559b6-tpnl9\" (UID: \"e1407e69-2431-4fdf-b8a7-b6396b46b6d9\") " pod="openshift-controller-manager/controller-manager-54464559b6-tpnl9" Feb 02 15:20:42 crc kubenswrapper[4733]: I0202 15:20:42.556336 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5cf5674648-xw84g" podStartSLOduration=4.556315725 podStartE2EDuration="4.556315725s" podCreationTimestamp="2026-02-02 15:20:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:20:42.553145885 +0000 UTC m=+326.004607253" watchObservedRunningTime="2026-02-02 15:20:42.556315725 +0000 UTC m=+326.007777093" Feb 02 15:20:42 crc kubenswrapper[4733]: I0202 15:20:42.556455 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e1407e69-2431-4fdf-b8a7-b6396b46b6d9-serving-cert\") pod \"controller-manager-54464559b6-tpnl9\" (UID: \"e1407e69-2431-4fdf-b8a7-b6396b46b6d9\") " pod="openshift-controller-manager/controller-manager-54464559b6-tpnl9" Feb 02 15:20:42 crc kubenswrapper[4733]: I0202 15:20:42.569329 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nj2s4\" (UniqueName: \"kubernetes.io/projected/e1407e69-2431-4fdf-b8a7-b6396b46b6d9-kube-api-access-nj2s4\") pod \"controller-manager-54464559b6-tpnl9\" (UID: \"e1407e69-2431-4fdf-b8a7-b6396b46b6d9\") " pod="openshift-controller-manager/controller-manager-54464559b6-tpnl9" Feb 02 15:20:42 crc kubenswrapper[4733]: I0202 15:20:42.612216 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-86cb48757f-sfmxz"] Feb 02 15:20:42 crc kubenswrapper[4733]: I0202 15:20:42.620552 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-86cb48757f-sfmxz"] Feb 02 15:20:42 crc kubenswrapper[4733]: I0202 15:20:42.704883 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-54464559b6-tpnl9" Feb 02 15:20:43 crc kubenswrapper[4733]: I0202 15:20:43.163572 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-54464559b6-tpnl9"] Feb 02 15:20:43 crc kubenswrapper[4733]: I0202 15:20:43.261321 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c0443cb-4d5e-4389-a935-1821f94f97a3" path="/var/lib/kubelet/pods/2c0443cb-4d5e-4389-a935-1821f94f97a3/volumes" Feb 02 15:20:43 crc kubenswrapper[4733]: I0202 15:20:43.542329 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-54464559b6-tpnl9" event={"ID":"e1407e69-2431-4fdf-b8a7-b6396b46b6d9","Type":"ContainerStarted","Data":"6747c0d370cf58da2862f3779f20a77bc346d0aaaf4784f3e5c73432169736b7"} Feb 02 15:20:43 crc kubenswrapper[4733]: I0202 15:20:43.542439 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-54464559b6-tpnl9" event={"ID":"e1407e69-2431-4fdf-b8a7-b6396b46b6d9","Type":"ContainerStarted","Data":"56ad4feb0f6070978a7683f08cb2b131edb0adc8841dc42500ac1796c06f6d4f"} Feb 02 15:20:43 crc kubenswrapper[4733]: I0202 15:20:43.542616 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-54464559b6-tpnl9" Feb 02 15:20:43 crc kubenswrapper[4733]: I0202 15:20:43.574705 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-54464559b6-tpnl9" podStartSLOduration=5.574678154 podStartE2EDuration="5.574678154s" podCreationTimestamp="2026-02-02 15:20:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:20:43.572348538 +0000 UTC m=+327.023809896" watchObservedRunningTime="2026-02-02 15:20:43.574678154 +0000 UTC m=+327.026139552" Feb 02 15:20:43 crc kubenswrapper[4733]: I0202 15:20:43.581996 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-54464559b6-tpnl9" Feb 02 15:21:12 crc kubenswrapper[4733]: I0202 15:21:12.451538 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5cf5674648-xw84g"] Feb 02 15:21:12 crc kubenswrapper[4733]: I0202 15:21:12.452127 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-5cf5674648-xw84g" podUID="ba456f3a-c060-4c07-b370-521e08c4f57a" containerName="route-controller-manager" containerID="cri-o://4abfe5d8edf004c3326af0a64d1160b23ef5ce149e733ee50e3aa121fb1585f2" gracePeriod=30 Feb 02 15:21:12 crc kubenswrapper[4733]: I0202 15:21:12.726154 4733 generic.go:334] "Generic (PLEG): container finished" podID="ba456f3a-c060-4c07-b370-521e08c4f57a" containerID="4abfe5d8edf004c3326af0a64d1160b23ef5ce149e733ee50e3aa121fb1585f2" exitCode=0 Feb 02 15:21:12 crc kubenswrapper[4733]: I0202 15:21:12.726461 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5cf5674648-xw84g" event={"ID":"ba456f3a-c060-4c07-b370-521e08c4f57a","Type":"ContainerDied","Data":"4abfe5d8edf004c3326af0a64d1160b23ef5ce149e733ee50e3aa121fb1585f2"} Feb 02 15:21:12 crc kubenswrapper[4733]: I0202 15:21:12.916860 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5cf5674648-xw84g" Feb 02 15:21:12 crc kubenswrapper[4733]: I0202 15:21:12.980221 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-574b2\" (UniqueName: \"kubernetes.io/projected/ba456f3a-c060-4c07-b370-521e08c4f57a-kube-api-access-574b2\") pod \"ba456f3a-c060-4c07-b370-521e08c4f57a\" (UID: \"ba456f3a-c060-4c07-b370-521e08c4f57a\") " Feb 02 15:21:12 crc kubenswrapper[4733]: I0202 15:21:12.980480 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba456f3a-c060-4c07-b370-521e08c4f57a-config\") pod \"ba456f3a-c060-4c07-b370-521e08c4f57a\" (UID: \"ba456f3a-c060-4c07-b370-521e08c4f57a\") " Feb 02 15:21:12 crc kubenswrapper[4733]: I0202 15:21:12.980592 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ba456f3a-c060-4c07-b370-521e08c4f57a-client-ca\") pod \"ba456f3a-c060-4c07-b370-521e08c4f57a\" (UID: \"ba456f3a-c060-4c07-b370-521e08c4f57a\") " Feb 02 15:21:12 crc kubenswrapper[4733]: I0202 15:21:12.980656 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ba456f3a-c060-4c07-b370-521e08c4f57a-serving-cert\") pod \"ba456f3a-c060-4c07-b370-521e08c4f57a\" (UID: \"ba456f3a-c060-4c07-b370-521e08c4f57a\") " Feb 02 15:21:12 crc kubenswrapper[4733]: I0202 15:21:12.982922 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba456f3a-c060-4c07-b370-521e08c4f57a-client-ca" (OuterVolumeSpecName: "client-ca") pod "ba456f3a-c060-4c07-b370-521e08c4f57a" (UID: "ba456f3a-c060-4c07-b370-521e08c4f57a"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:21:12 crc kubenswrapper[4733]: I0202 15:21:12.982999 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba456f3a-c060-4c07-b370-521e08c4f57a-config" (OuterVolumeSpecName: "config") pod "ba456f3a-c060-4c07-b370-521e08c4f57a" (UID: "ba456f3a-c060-4c07-b370-521e08c4f57a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:21:12 crc kubenswrapper[4733]: I0202 15:21:12.988259 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba456f3a-c060-4c07-b370-521e08c4f57a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "ba456f3a-c060-4c07-b370-521e08c4f57a" (UID: "ba456f3a-c060-4c07-b370-521e08c4f57a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:21:12 crc kubenswrapper[4733]: I0202 15:21:12.988268 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba456f3a-c060-4c07-b370-521e08c4f57a-kube-api-access-574b2" (OuterVolumeSpecName: "kube-api-access-574b2") pod "ba456f3a-c060-4c07-b370-521e08c4f57a" (UID: "ba456f3a-c060-4c07-b370-521e08c4f57a"). InnerVolumeSpecName "kube-api-access-574b2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:21:13 crc kubenswrapper[4733]: I0202 15:21:13.082087 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ba456f3a-c060-4c07-b370-521e08c4f57a-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 15:21:13 crc kubenswrapper[4733]: I0202 15:21:13.082133 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-574b2\" (UniqueName: \"kubernetes.io/projected/ba456f3a-c060-4c07-b370-521e08c4f57a-kube-api-access-574b2\") on node \"crc\" DevicePath \"\"" Feb 02 15:21:13 crc kubenswrapper[4733]: I0202 15:21:13.082152 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba456f3a-c060-4c07-b370-521e08c4f57a-config\") on node \"crc\" DevicePath \"\"" Feb 02 15:21:13 crc kubenswrapper[4733]: I0202 15:21:13.082192 4733 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ba456f3a-c060-4c07-b370-521e08c4f57a-client-ca\") on node \"crc\" DevicePath \"\"" Feb 02 15:21:13 crc kubenswrapper[4733]: I0202 15:21:13.735459 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5cf5674648-xw84g" event={"ID":"ba456f3a-c060-4c07-b370-521e08c4f57a","Type":"ContainerDied","Data":"c581c02249b43aed6bbfdb1458f38591321224b89ff14be0ab7a4f41795960b5"} Feb 02 15:21:13 crc kubenswrapper[4733]: I0202 15:21:13.735801 4733 scope.go:117] "RemoveContainer" containerID="4abfe5d8edf004c3326af0a64d1160b23ef5ce149e733ee50e3aa121fb1585f2" Feb 02 15:21:13 crc kubenswrapper[4733]: I0202 15:21:13.735526 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5cf5674648-xw84g" Feb 02 15:21:13 crc kubenswrapper[4733]: I0202 15:21:13.761122 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5cf5674648-xw84g"] Feb 02 15:21:13 crc kubenswrapper[4733]: I0202 15:21:13.769984 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5cf5674648-xw84g"] Feb 02 15:21:13 crc kubenswrapper[4733]: I0202 15:21:13.832867 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-747d45d6f5-q6h8z"] Feb 02 15:21:13 crc kubenswrapper[4733]: E0202 15:21:13.833413 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba456f3a-c060-4c07-b370-521e08c4f57a" containerName="route-controller-manager" Feb 02 15:21:13 crc kubenswrapper[4733]: I0202 15:21:13.833497 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba456f3a-c060-4c07-b370-521e08c4f57a" containerName="route-controller-manager" Feb 02 15:21:13 crc kubenswrapper[4733]: I0202 15:21:13.834074 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba456f3a-c060-4c07-b370-521e08c4f57a" containerName="route-controller-manager" Feb 02 15:21:13 crc kubenswrapper[4733]: I0202 15:21:13.835095 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-747d45d6f5-q6h8z" Feb 02 15:21:13 crc kubenswrapper[4733]: I0202 15:21:13.840484 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 02 15:21:13 crc kubenswrapper[4733]: I0202 15:21:13.840586 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 02 15:21:13 crc kubenswrapper[4733]: I0202 15:21:13.840646 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 02 15:21:13 crc kubenswrapper[4733]: I0202 15:21:13.840708 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 02 15:21:13 crc kubenswrapper[4733]: I0202 15:21:13.840745 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 02 15:21:13 crc kubenswrapper[4733]: I0202 15:21:13.840785 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 02 15:21:13 crc kubenswrapper[4733]: I0202 15:21:13.848681 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-747d45d6f5-q6h8z"] Feb 02 15:21:13 crc kubenswrapper[4733]: I0202 15:21:13.893366 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrgd2\" (UniqueName: \"kubernetes.io/projected/ddb3f050-cd63-44d3-a759-63f061f72b8a-kube-api-access-wrgd2\") pod \"route-controller-manager-747d45d6f5-q6h8z\" (UID: \"ddb3f050-cd63-44d3-a759-63f061f72b8a\") " pod="openshift-route-controller-manager/route-controller-manager-747d45d6f5-q6h8z" Feb 02 15:21:13 crc kubenswrapper[4733]: I0202 15:21:13.893426 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ddb3f050-cd63-44d3-a759-63f061f72b8a-serving-cert\") pod \"route-controller-manager-747d45d6f5-q6h8z\" (UID: \"ddb3f050-cd63-44d3-a759-63f061f72b8a\") " pod="openshift-route-controller-manager/route-controller-manager-747d45d6f5-q6h8z" Feb 02 15:21:13 crc kubenswrapper[4733]: I0202 15:21:13.893461 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ddb3f050-cd63-44d3-a759-63f061f72b8a-client-ca\") pod \"route-controller-manager-747d45d6f5-q6h8z\" (UID: \"ddb3f050-cd63-44d3-a759-63f061f72b8a\") " pod="openshift-route-controller-manager/route-controller-manager-747d45d6f5-q6h8z" Feb 02 15:21:13 crc kubenswrapper[4733]: I0202 15:21:13.893543 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ddb3f050-cd63-44d3-a759-63f061f72b8a-config\") pod \"route-controller-manager-747d45d6f5-q6h8z\" (UID: \"ddb3f050-cd63-44d3-a759-63f061f72b8a\") " pod="openshift-route-controller-manager/route-controller-manager-747d45d6f5-q6h8z" Feb 02 15:21:13 crc kubenswrapper[4733]: I0202 15:21:13.994847 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ddb3f050-cd63-44d3-a759-63f061f72b8a-client-ca\") pod \"route-controller-manager-747d45d6f5-q6h8z\" (UID: \"ddb3f050-cd63-44d3-a759-63f061f72b8a\") " pod="openshift-route-controller-manager/route-controller-manager-747d45d6f5-q6h8z" Feb 02 15:21:13 crc kubenswrapper[4733]: I0202 15:21:13.994926 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ddb3f050-cd63-44d3-a759-63f061f72b8a-config\") pod \"route-controller-manager-747d45d6f5-q6h8z\" (UID: \"ddb3f050-cd63-44d3-a759-63f061f72b8a\") " pod="openshift-route-controller-manager/route-controller-manager-747d45d6f5-q6h8z" Feb 02 15:21:13 crc kubenswrapper[4733]: I0202 15:21:13.995033 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrgd2\" (UniqueName: \"kubernetes.io/projected/ddb3f050-cd63-44d3-a759-63f061f72b8a-kube-api-access-wrgd2\") pod \"route-controller-manager-747d45d6f5-q6h8z\" (UID: \"ddb3f050-cd63-44d3-a759-63f061f72b8a\") " pod="openshift-route-controller-manager/route-controller-manager-747d45d6f5-q6h8z" Feb 02 15:21:13 crc kubenswrapper[4733]: I0202 15:21:13.995083 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ddb3f050-cd63-44d3-a759-63f061f72b8a-serving-cert\") pod \"route-controller-manager-747d45d6f5-q6h8z\" (UID: \"ddb3f050-cd63-44d3-a759-63f061f72b8a\") " pod="openshift-route-controller-manager/route-controller-manager-747d45d6f5-q6h8z" Feb 02 15:21:13 crc kubenswrapper[4733]: I0202 15:21:13.996966 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ddb3f050-cd63-44d3-a759-63f061f72b8a-client-ca\") pod \"route-controller-manager-747d45d6f5-q6h8z\" (UID: \"ddb3f050-cd63-44d3-a759-63f061f72b8a\") " pod="openshift-route-controller-manager/route-controller-manager-747d45d6f5-q6h8z" Feb 02 15:21:13 crc kubenswrapper[4733]: I0202 15:21:13.997309 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ddb3f050-cd63-44d3-a759-63f061f72b8a-config\") pod \"route-controller-manager-747d45d6f5-q6h8z\" (UID: \"ddb3f050-cd63-44d3-a759-63f061f72b8a\") " pod="openshift-route-controller-manager/route-controller-manager-747d45d6f5-q6h8z" Feb 02 15:21:14 crc kubenswrapper[4733]: I0202 15:21:14.006465 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ddb3f050-cd63-44d3-a759-63f061f72b8a-serving-cert\") pod \"route-controller-manager-747d45d6f5-q6h8z\" (UID: \"ddb3f050-cd63-44d3-a759-63f061f72b8a\") " pod="openshift-route-controller-manager/route-controller-manager-747d45d6f5-q6h8z" Feb 02 15:21:14 crc kubenswrapper[4733]: I0202 15:21:14.030612 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrgd2\" (UniqueName: \"kubernetes.io/projected/ddb3f050-cd63-44d3-a759-63f061f72b8a-kube-api-access-wrgd2\") pod \"route-controller-manager-747d45d6f5-q6h8z\" (UID: \"ddb3f050-cd63-44d3-a759-63f061f72b8a\") " pod="openshift-route-controller-manager/route-controller-manager-747d45d6f5-q6h8z" Feb 02 15:21:14 crc kubenswrapper[4733]: I0202 15:21:14.149694 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-747d45d6f5-q6h8z" Feb 02 15:21:14 crc kubenswrapper[4733]: I0202 15:21:14.689107 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-747d45d6f5-q6h8z"] Feb 02 15:21:14 crc kubenswrapper[4733]: I0202 15:21:14.745298 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-747d45d6f5-q6h8z" event={"ID":"ddb3f050-cd63-44d3-a759-63f061f72b8a","Type":"ContainerStarted","Data":"5fb0f482557cdbb34e3c5a4465fb359c4fe937b70ee9ca97e171751f3b4bf652"} Feb 02 15:21:15 crc kubenswrapper[4733]: I0202 15:21:15.265969 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba456f3a-c060-4c07-b370-521e08c4f57a" path="/var/lib/kubelet/pods/ba456f3a-c060-4c07-b370-521e08c4f57a/volumes" Feb 02 15:21:15 crc kubenswrapper[4733]: I0202 15:21:15.483307 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-brfn5"] Feb 02 15:21:15 crc kubenswrapper[4733]: I0202 15:21:15.484047 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-brfn5" Feb 02 15:21:15 crc kubenswrapper[4733]: I0202 15:21:15.511612 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-brfn5"] Feb 02 15:21:15 crc kubenswrapper[4733]: I0202 15:21:15.543470 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/dfb81b50-dd41-485e-ac5d-c4bd89d7b9ba-registry-tls\") pod \"image-registry-66df7c8f76-brfn5\" (UID: \"dfb81b50-dd41-485e-ac5d-c4bd89d7b9ba\") " pod="openshift-image-registry/image-registry-66df7c8f76-brfn5" Feb 02 15:21:15 crc kubenswrapper[4733]: I0202 15:21:15.543526 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dfb81b50-dd41-485e-ac5d-c4bd89d7b9ba-trusted-ca\") pod \"image-registry-66df7c8f76-brfn5\" (UID: \"dfb81b50-dd41-485e-ac5d-c4bd89d7b9ba\") " pod="openshift-image-registry/image-registry-66df7c8f76-brfn5" Feb 02 15:21:15 crc kubenswrapper[4733]: I0202 15:21:15.543584 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/dfb81b50-dd41-485e-ac5d-c4bd89d7b9ba-bound-sa-token\") pod \"image-registry-66df7c8f76-brfn5\" (UID: \"dfb81b50-dd41-485e-ac5d-c4bd89d7b9ba\") " pod="openshift-image-registry/image-registry-66df7c8f76-brfn5" Feb 02 15:21:15 crc kubenswrapper[4733]: I0202 15:21:15.543606 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/dfb81b50-dd41-485e-ac5d-c4bd89d7b9ba-ca-trust-extracted\") pod \"image-registry-66df7c8f76-brfn5\" (UID: \"dfb81b50-dd41-485e-ac5d-c4bd89d7b9ba\") " pod="openshift-image-registry/image-registry-66df7c8f76-brfn5" Feb 02 15:21:15 crc kubenswrapper[4733]: I0202 15:21:15.543629 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsldr\" (UniqueName: \"kubernetes.io/projected/dfb81b50-dd41-485e-ac5d-c4bd89d7b9ba-kube-api-access-dsldr\") pod \"image-registry-66df7c8f76-brfn5\" (UID: \"dfb81b50-dd41-485e-ac5d-c4bd89d7b9ba\") " pod="openshift-image-registry/image-registry-66df7c8f76-brfn5" Feb 02 15:21:15 crc kubenswrapper[4733]: I0202 15:21:15.543670 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-brfn5\" (UID: \"dfb81b50-dd41-485e-ac5d-c4bd89d7b9ba\") " pod="openshift-image-registry/image-registry-66df7c8f76-brfn5" Feb 02 15:21:15 crc kubenswrapper[4733]: I0202 15:21:15.543703 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/dfb81b50-dd41-485e-ac5d-c4bd89d7b9ba-registry-certificates\") pod \"image-registry-66df7c8f76-brfn5\" (UID: \"dfb81b50-dd41-485e-ac5d-c4bd89d7b9ba\") " pod="openshift-image-registry/image-registry-66df7c8f76-brfn5" Feb 02 15:21:15 crc kubenswrapper[4733]: I0202 15:21:15.543757 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/dfb81b50-dd41-485e-ac5d-c4bd89d7b9ba-installation-pull-secrets\") pod \"image-registry-66df7c8f76-brfn5\" (UID: \"dfb81b50-dd41-485e-ac5d-c4bd89d7b9ba\") " pod="openshift-image-registry/image-registry-66df7c8f76-brfn5" Feb 02 15:21:15 crc kubenswrapper[4733]: I0202 15:21:15.564684 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-brfn5\" (UID: \"dfb81b50-dd41-485e-ac5d-c4bd89d7b9ba\") " pod="openshift-image-registry/image-registry-66df7c8f76-brfn5" Feb 02 15:21:15 crc kubenswrapper[4733]: I0202 15:21:15.645213 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/dfb81b50-dd41-485e-ac5d-c4bd89d7b9ba-bound-sa-token\") pod \"image-registry-66df7c8f76-brfn5\" (UID: \"dfb81b50-dd41-485e-ac5d-c4bd89d7b9ba\") " pod="openshift-image-registry/image-registry-66df7c8f76-brfn5" Feb 02 15:21:15 crc kubenswrapper[4733]: I0202 15:21:15.645261 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/dfb81b50-dd41-485e-ac5d-c4bd89d7b9ba-ca-trust-extracted\") pod \"image-registry-66df7c8f76-brfn5\" (UID: \"dfb81b50-dd41-485e-ac5d-c4bd89d7b9ba\") " pod="openshift-image-registry/image-registry-66df7c8f76-brfn5" Feb 02 15:21:15 crc kubenswrapper[4733]: I0202 15:21:15.645284 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dsldr\" (UniqueName: \"kubernetes.io/projected/dfb81b50-dd41-485e-ac5d-c4bd89d7b9ba-kube-api-access-dsldr\") pod \"image-registry-66df7c8f76-brfn5\" (UID: \"dfb81b50-dd41-485e-ac5d-c4bd89d7b9ba\") " pod="openshift-image-registry/image-registry-66df7c8f76-brfn5" Feb 02 15:21:15 crc kubenswrapper[4733]: I0202 15:21:15.645311 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/dfb81b50-dd41-485e-ac5d-c4bd89d7b9ba-registry-certificates\") pod \"image-registry-66df7c8f76-brfn5\" (UID: \"dfb81b50-dd41-485e-ac5d-c4bd89d7b9ba\") " pod="openshift-image-registry/image-registry-66df7c8f76-brfn5" Feb 02 15:21:15 crc kubenswrapper[4733]: I0202 15:21:15.645344 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/dfb81b50-dd41-485e-ac5d-c4bd89d7b9ba-installation-pull-secrets\") pod \"image-registry-66df7c8f76-brfn5\" (UID: \"dfb81b50-dd41-485e-ac5d-c4bd89d7b9ba\") " pod="openshift-image-registry/image-registry-66df7c8f76-brfn5" Feb 02 15:21:15 crc kubenswrapper[4733]: I0202 15:21:15.645377 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/dfb81b50-dd41-485e-ac5d-c4bd89d7b9ba-registry-tls\") pod \"image-registry-66df7c8f76-brfn5\" (UID: \"dfb81b50-dd41-485e-ac5d-c4bd89d7b9ba\") " pod="openshift-image-registry/image-registry-66df7c8f76-brfn5" Feb 02 15:21:15 crc kubenswrapper[4733]: I0202 15:21:15.645394 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dfb81b50-dd41-485e-ac5d-c4bd89d7b9ba-trusted-ca\") pod \"image-registry-66df7c8f76-brfn5\" (UID: \"dfb81b50-dd41-485e-ac5d-c4bd89d7b9ba\") " pod="openshift-image-registry/image-registry-66df7c8f76-brfn5" Feb 02 15:21:15 crc kubenswrapper[4733]: I0202 15:21:15.646594 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dfb81b50-dd41-485e-ac5d-c4bd89d7b9ba-trusted-ca\") pod \"image-registry-66df7c8f76-brfn5\" (UID: \"dfb81b50-dd41-485e-ac5d-c4bd89d7b9ba\") " pod="openshift-image-registry/image-registry-66df7c8f76-brfn5" Feb 02 15:21:15 crc kubenswrapper[4733]: I0202 15:21:15.646746 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/dfb81b50-dd41-485e-ac5d-c4bd89d7b9ba-ca-trust-extracted\") pod \"image-registry-66df7c8f76-brfn5\" (UID: \"dfb81b50-dd41-485e-ac5d-c4bd89d7b9ba\") " pod="openshift-image-registry/image-registry-66df7c8f76-brfn5" Feb 02 15:21:15 crc kubenswrapper[4733]: I0202 15:21:15.646842 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/dfb81b50-dd41-485e-ac5d-c4bd89d7b9ba-registry-certificates\") pod \"image-registry-66df7c8f76-brfn5\" (UID: \"dfb81b50-dd41-485e-ac5d-c4bd89d7b9ba\") " pod="openshift-image-registry/image-registry-66df7c8f76-brfn5" Feb 02 15:21:15 crc kubenswrapper[4733]: I0202 15:21:15.651405 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/dfb81b50-dd41-485e-ac5d-c4bd89d7b9ba-registry-tls\") pod \"image-registry-66df7c8f76-brfn5\" (UID: \"dfb81b50-dd41-485e-ac5d-c4bd89d7b9ba\") " pod="openshift-image-registry/image-registry-66df7c8f76-brfn5" Feb 02 15:21:15 crc kubenswrapper[4733]: I0202 15:21:15.657551 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/dfb81b50-dd41-485e-ac5d-c4bd89d7b9ba-installation-pull-secrets\") pod \"image-registry-66df7c8f76-brfn5\" (UID: \"dfb81b50-dd41-485e-ac5d-c4bd89d7b9ba\") " pod="openshift-image-registry/image-registry-66df7c8f76-brfn5" Feb 02 15:21:15 crc kubenswrapper[4733]: I0202 15:21:15.663528 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsldr\" (UniqueName: \"kubernetes.io/projected/dfb81b50-dd41-485e-ac5d-c4bd89d7b9ba-kube-api-access-dsldr\") pod \"image-registry-66df7c8f76-brfn5\" (UID: \"dfb81b50-dd41-485e-ac5d-c4bd89d7b9ba\") " pod="openshift-image-registry/image-registry-66df7c8f76-brfn5" Feb 02 15:21:15 crc kubenswrapper[4733]: I0202 15:21:15.667143 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/dfb81b50-dd41-485e-ac5d-c4bd89d7b9ba-bound-sa-token\") pod \"image-registry-66df7c8f76-brfn5\" (UID: \"dfb81b50-dd41-485e-ac5d-c4bd89d7b9ba\") " pod="openshift-image-registry/image-registry-66df7c8f76-brfn5" Feb 02 15:21:15 crc kubenswrapper[4733]: I0202 15:21:15.764306 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-747d45d6f5-q6h8z" event={"ID":"ddb3f050-cd63-44d3-a759-63f061f72b8a","Type":"ContainerStarted","Data":"2be3f29572c309fee522ed5ad46cdfb9e97fadf1c59cddb89f07eb96ec05971d"} Feb 02 15:21:15 crc kubenswrapper[4733]: I0202 15:21:15.764754 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-747d45d6f5-q6h8z" Feb 02 15:21:15 crc kubenswrapper[4733]: I0202 15:21:15.772148 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-747d45d6f5-q6h8z" Feb 02 15:21:15 crc kubenswrapper[4733]: I0202 15:21:15.784043 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-747d45d6f5-q6h8z" podStartSLOduration=3.784025347 podStartE2EDuration="3.784025347s" podCreationTimestamp="2026-02-02 15:21:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:21:15.783382899 +0000 UTC m=+359.234844267" watchObservedRunningTime="2026-02-02 15:21:15.784025347 +0000 UTC m=+359.235486705" Feb 02 15:21:15 crc kubenswrapper[4733]: I0202 15:21:15.799417 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-brfn5" Feb 02 15:21:16 crc kubenswrapper[4733]: I0202 15:21:16.223696 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-brfn5"] Feb 02 15:21:16 crc kubenswrapper[4733]: W0202 15:21:16.234141 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddfb81b50_dd41_485e_ac5d_c4bd89d7b9ba.slice/crio-db84124efb566a1e16cc3d139864b08b94ac5806018093ce99b5daaea49ac902 WatchSource:0}: Error finding container db84124efb566a1e16cc3d139864b08b94ac5806018093ce99b5daaea49ac902: Status 404 returned error can't find the container with id db84124efb566a1e16cc3d139864b08b94ac5806018093ce99b5daaea49ac902 Feb 02 15:21:16 crc kubenswrapper[4733]: I0202 15:21:16.773726 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-brfn5" event={"ID":"dfb81b50-dd41-485e-ac5d-c4bd89d7b9ba","Type":"ContainerStarted","Data":"d8b60d6a55af96c7f451f4f34d107158dde8fea77771674e820b4a94c8521658"} Feb 02 15:21:16 crc kubenswrapper[4733]: I0202 15:21:16.774114 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-brfn5" event={"ID":"dfb81b50-dd41-485e-ac5d-c4bd89d7b9ba","Type":"ContainerStarted","Data":"db84124efb566a1e16cc3d139864b08b94ac5806018093ce99b5daaea49ac902"} Feb 02 15:21:16 crc kubenswrapper[4733]: I0202 15:21:16.774199 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-brfn5" Feb 02 15:21:16 crc kubenswrapper[4733]: I0202 15:21:16.806131 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-brfn5" podStartSLOduration=1.806108622 podStartE2EDuration="1.806108622s" podCreationTimestamp="2026-02-02 15:21:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:21:16.800592315 +0000 UTC m=+360.252053693" watchObservedRunningTime="2026-02-02 15:21:16.806108622 +0000 UTC m=+360.257570000" Feb 02 15:21:19 crc kubenswrapper[4733]: I0202 15:21:19.406457 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7gm8t"] Feb 02 15:21:19 crc kubenswrapper[4733]: I0202 15:21:19.408419 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7gm8t" podUID="d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2" containerName="registry-server" containerID="cri-o://720a4dabd902a96d042261a42019bb9ea3beee89af69a1d942baecce6063882e" gracePeriod=30 Feb 02 15:21:19 crc kubenswrapper[4733]: I0202 15:21:19.424714 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jk5b2"] Feb 02 15:21:19 crc kubenswrapper[4733]: I0202 15:21:19.440508 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-csdnv"] Feb 02 15:21:19 crc kubenswrapper[4733]: I0202 15:21:19.440835 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-csdnv" podUID="ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e" containerName="marketplace-operator" containerID="cri-o://35db594c74e3ce27d2532036e1ab3f1815c9fed37332b6cab1836020121772db" gracePeriod=30 Feb 02 15:21:19 crc kubenswrapper[4733]: I0202 15:21:19.471879 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-v5t9s"] Feb 02 15:21:19 crc kubenswrapper[4733]: I0202 15:21:19.472382 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-v5t9s" podUID="7b6ed8e1-1992-4d83-954a-552aa8e7acb3" containerName="registry-server" containerID="cri-o://c9aa09ab01768bd69cd858d98490db679284ae40c82f880a58ece8b87c424d3f" gracePeriod=30 Feb 02 15:21:19 crc kubenswrapper[4733]: I0202 15:21:19.480830 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ps6dr"] Feb 02 15:21:19 crc kubenswrapper[4733]: I0202 15:21:19.482033 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ps6dr" Feb 02 15:21:19 crc kubenswrapper[4733]: I0202 15:21:19.485581 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9fckn"] Feb 02 15:21:19 crc kubenswrapper[4733]: I0202 15:21:19.485855 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9fckn" podUID="52a641c3-5baf-4d1d-a51f-93f372b1013e" containerName="registry-server" containerID="cri-o://bc16b500617a150f81ab1c82995e4be4e445957acddb58d55773c44c85ff85e7" gracePeriod=30 Feb 02 15:21:19 crc kubenswrapper[4733]: I0202 15:21:19.493291 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ps6dr"] Feb 02 15:21:19 crc kubenswrapper[4733]: I0202 15:21:19.510137 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/babbc7b3-6e09-43a7-9cbc-70ba7076ab9f-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ps6dr\" (UID: \"babbc7b3-6e09-43a7-9cbc-70ba7076ab9f\") " pod="openshift-marketplace/marketplace-operator-79b997595-ps6dr" Feb 02 15:21:19 crc kubenswrapper[4733]: I0202 15:21:19.510473 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cl6zg\" (UniqueName: \"kubernetes.io/projected/babbc7b3-6e09-43a7-9cbc-70ba7076ab9f-kube-api-access-cl6zg\") pod \"marketplace-operator-79b997595-ps6dr\" (UID: \"babbc7b3-6e09-43a7-9cbc-70ba7076ab9f\") " pod="openshift-marketplace/marketplace-operator-79b997595-ps6dr" Feb 02 15:21:19 crc kubenswrapper[4733]: I0202 15:21:19.510617 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/babbc7b3-6e09-43a7-9cbc-70ba7076ab9f-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ps6dr\" (UID: \"babbc7b3-6e09-43a7-9cbc-70ba7076ab9f\") " pod="openshift-marketplace/marketplace-operator-79b997595-ps6dr" Feb 02 15:21:19 crc kubenswrapper[4733]: I0202 15:21:19.621618 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cl6zg\" (UniqueName: \"kubernetes.io/projected/babbc7b3-6e09-43a7-9cbc-70ba7076ab9f-kube-api-access-cl6zg\") pod \"marketplace-operator-79b997595-ps6dr\" (UID: \"babbc7b3-6e09-43a7-9cbc-70ba7076ab9f\") " pod="openshift-marketplace/marketplace-operator-79b997595-ps6dr" Feb 02 15:21:19 crc kubenswrapper[4733]: I0202 15:21:19.621690 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/babbc7b3-6e09-43a7-9cbc-70ba7076ab9f-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ps6dr\" (UID: \"babbc7b3-6e09-43a7-9cbc-70ba7076ab9f\") " pod="openshift-marketplace/marketplace-operator-79b997595-ps6dr" Feb 02 15:21:19 crc kubenswrapper[4733]: I0202 15:21:19.621742 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/babbc7b3-6e09-43a7-9cbc-70ba7076ab9f-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ps6dr\" (UID: \"babbc7b3-6e09-43a7-9cbc-70ba7076ab9f\") " pod="openshift-marketplace/marketplace-operator-79b997595-ps6dr" Feb 02 15:21:19 crc kubenswrapper[4733]: I0202 15:21:19.635272 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/babbc7b3-6e09-43a7-9cbc-70ba7076ab9f-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ps6dr\" (UID: \"babbc7b3-6e09-43a7-9cbc-70ba7076ab9f\") " pod="openshift-marketplace/marketplace-operator-79b997595-ps6dr" Feb 02 15:21:19 crc kubenswrapper[4733]: I0202 15:21:19.635511 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/babbc7b3-6e09-43a7-9cbc-70ba7076ab9f-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ps6dr\" (UID: \"babbc7b3-6e09-43a7-9cbc-70ba7076ab9f\") " pod="openshift-marketplace/marketplace-operator-79b997595-ps6dr" Feb 02 15:21:19 crc kubenswrapper[4733]: I0202 15:21:19.639783 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cl6zg\" (UniqueName: \"kubernetes.io/projected/babbc7b3-6e09-43a7-9cbc-70ba7076ab9f-kube-api-access-cl6zg\") pod \"marketplace-operator-79b997595-ps6dr\" (UID: \"babbc7b3-6e09-43a7-9cbc-70ba7076ab9f\") " pod="openshift-marketplace/marketplace-operator-79b997595-ps6dr" Feb 02 15:21:19 crc kubenswrapper[4733]: I0202 15:21:19.792062 4733 generic.go:334] "Generic (PLEG): container finished" podID="52a641c3-5baf-4d1d-a51f-93f372b1013e" containerID="bc16b500617a150f81ab1c82995e4be4e445957acddb58d55773c44c85ff85e7" exitCode=0 Feb 02 15:21:19 crc kubenswrapper[4733]: I0202 15:21:19.792815 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9fckn" event={"ID":"52a641c3-5baf-4d1d-a51f-93f372b1013e","Type":"ContainerDied","Data":"bc16b500617a150f81ab1c82995e4be4e445957acddb58d55773c44c85ff85e7"} Feb 02 15:21:19 crc kubenswrapper[4733]: I0202 15:21:19.796531 4733 generic.go:334] "Generic (PLEG): container finished" podID="7b6ed8e1-1992-4d83-954a-552aa8e7acb3" containerID="c9aa09ab01768bd69cd858d98490db679284ae40c82f880a58ece8b87c424d3f" exitCode=0 Feb 02 15:21:19 crc kubenswrapper[4733]: I0202 15:21:19.796624 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v5t9s" event={"ID":"7b6ed8e1-1992-4d83-954a-552aa8e7acb3","Type":"ContainerDied","Data":"c9aa09ab01768bd69cd858d98490db679284ae40c82f880a58ece8b87c424d3f"} Feb 02 15:21:19 crc kubenswrapper[4733]: I0202 15:21:19.798687 4733 generic.go:334] "Generic (PLEG): container finished" podID="d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2" containerID="720a4dabd902a96d042261a42019bb9ea3beee89af69a1d942baecce6063882e" exitCode=0 Feb 02 15:21:19 crc kubenswrapper[4733]: I0202 15:21:19.798763 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7gm8t" event={"ID":"d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2","Type":"ContainerDied","Data":"720a4dabd902a96d042261a42019bb9ea3beee89af69a1d942baecce6063882e"} Feb 02 15:21:19 crc kubenswrapper[4733]: I0202 15:21:19.804404 4733 generic.go:334] "Generic (PLEG): container finished" podID="ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e" containerID="35db594c74e3ce27d2532036e1ab3f1815c9fed37332b6cab1836020121772db" exitCode=0 Feb 02 15:21:19 crc kubenswrapper[4733]: I0202 15:21:19.804432 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-csdnv" event={"ID":"ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e","Type":"ContainerDied","Data":"35db594c74e3ce27d2532036e1ab3f1815c9fed37332b6cab1836020121772db"} Feb 02 15:21:19 crc kubenswrapper[4733]: I0202 15:21:19.804790 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jk5b2" podUID="4f03c333-e449-4a18-8efb-f8834911a599" containerName="registry-server" containerID="cri-o://e4a4c0a9fa82721f255c6e6572b720748b1dd22edba932206f6daaf366db2c53" gracePeriod=30 Feb 02 15:21:19 crc kubenswrapper[4733]: I0202 15:21:19.873395 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ps6dr" Feb 02 15:21:19 crc kubenswrapper[4733]: I0202 15:21:19.873541 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7gm8t" Feb 02 15:21:19 crc kubenswrapper[4733]: I0202 15:21:19.935825 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-csdnv" Feb 02 15:21:19 crc kubenswrapper[4733]: I0202 15:21:19.937944 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2-utilities\") pod \"d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2\" (UID: \"d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2\") " Feb 02 15:21:19 crc kubenswrapper[4733]: I0202 15:21:19.937994 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2-catalog-content\") pod \"d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2\" (UID: \"d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2\") " Feb 02 15:21:19 crc kubenswrapper[4733]: I0202 15:21:19.938047 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2kvxx\" (UniqueName: \"kubernetes.io/projected/d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2-kube-api-access-2kvxx\") pod \"d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2\" (UID: \"d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2\") " Feb 02 15:21:19 crc kubenswrapper[4733]: I0202 15:21:19.941005 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2-utilities" (OuterVolumeSpecName: "utilities") pod "d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2" (UID: "d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:21:19 crc kubenswrapper[4733]: I0202 15:21:19.943931 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2-kube-api-access-2kvxx" (OuterVolumeSpecName: "kube-api-access-2kvxx") pod "d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2" (UID: "d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2"). InnerVolumeSpecName "kube-api-access-2kvxx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:21:19 crc kubenswrapper[4733]: I0202 15:21:19.957749 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v5t9s" Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.011686 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2" (UID: "d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.013380 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9fckn" Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.039796 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e-marketplace-trusted-ca\") pod \"ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e\" (UID: \"ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e\") " Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.039869 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f7vzn\" (UniqueName: \"kubernetes.io/projected/7b6ed8e1-1992-4d83-954a-552aa8e7acb3-kube-api-access-f7vzn\") pod \"7b6ed8e1-1992-4d83-954a-552aa8e7acb3\" (UID: \"7b6ed8e1-1992-4d83-954a-552aa8e7acb3\") " Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.039948 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b6ed8e1-1992-4d83-954a-552aa8e7acb3-utilities\") pod \"7b6ed8e1-1992-4d83-954a-552aa8e7acb3\" (UID: \"7b6ed8e1-1992-4d83-954a-552aa8e7acb3\") " Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.039972 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7lvtx\" (UniqueName: \"kubernetes.io/projected/ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e-kube-api-access-7lvtx\") pod \"ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e\" (UID: \"ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e\") " Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.040006 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e-marketplace-operator-metrics\") pod \"ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e\" (UID: \"ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e\") " Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.040028 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b6ed8e1-1992-4d83-954a-552aa8e7acb3-catalog-content\") pod \"7b6ed8e1-1992-4d83-954a-552aa8e7acb3\" (UID: \"7b6ed8e1-1992-4d83-954a-552aa8e7acb3\") " Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.040324 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.040349 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.040366 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2kvxx\" (UniqueName: \"kubernetes.io/projected/d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2-kube-api-access-2kvxx\") on node \"crc\" DevicePath \"\"" Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.040372 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e" (UID: "ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.040829 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b6ed8e1-1992-4d83-954a-552aa8e7acb3-utilities" (OuterVolumeSpecName: "utilities") pod "7b6ed8e1-1992-4d83-954a-552aa8e7acb3" (UID: "7b6ed8e1-1992-4d83-954a-552aa8e7acb3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.045431 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b6ed8e1-1992-4d83-954a-552aa8e7acb3-kube-api-access-f7vzn" (OuterVolumeSpecName: "kube-api-access-f7vzn") pod "7b6ed8e1-1992-4d83-954a-552aa8e7acb3" (UID: "7b6ed8e1-1992-4d83-954a-552aa8e7acb3"). InnerVolumeSpecName "kube-api-access-f7vzn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.046398 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e-kube-api-access-7lvtx" (OuterVolumeSpecName: "kube-api-access-7lvtx") pod "ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e" (UID: "ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e"). InnerVolumeSpecName "kube-api-access-7lvtx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.050523 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e" (UID: "ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.063676 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b6ed8e1-1992-4d83-954a-552aa8e7acb3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7b6ed8e1-1992-4d83-954a-552aa8e7acb3" (UID: "7b6ed8e1-1992-4d83-954a-552aa8e7acb3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.140817 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52a641c3-5baf-4d1d-a51f-93f372b1013e-catalog-content\") pod \"52a641c3-5baf-4d1d-a51f-93f372b1013e\" (UID: \"52a641c3-5baf-4d1d-a51f-93f372b1013e\") " Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.140889 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q2fgr\" (UniqueName: \"kubernetes.io/projected/52a641c3-5baf-4d1d-a51f-93f372b1013e-kube-api-access-q2fgr\") pod \"52a641c3-5baf-4d1d-a51f-93f372b1013e\" (UID: \"52a641c3-5baf-4d1d-a51f-93f372b1013e\") " Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.140950 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52a641c3-5baf-4d1d-a51f-93f372b1013e-utilities\") pod \"52a641c3-5baf-4d1d-a51f-93f372b1013e\" (UID: \"52a641c3-5baf-4d1d-a51f-93f372b1013e\") " Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.141267 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f7vzn\" (UniqueName: \"kubernetes.io/projected/7b6ed8e1-1992-4d83-954a-552aa8e7acb3-kube-api-access-f7vzn\") on node \"crc\" DevicePath \"\"" Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.141286 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b6ed8e1-1992-4d83-954a-552aa8e7acb3-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.141298 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7lvtx\" (UniqueName: \"kubernetes.io/projected/ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e-kube-api-access-7lvtx\") on node \"crc\" DevicePath \"\"" Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.141309 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b6ed8e1-1992-4d83-954a-552aa8e7acb3-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.141322 4733 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.141334 4733 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.142538 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/52a641c3-5baf-4d1d-a51f-93f372b1013e-utilities" (OuterVolumeSpecName: "utilities") pod "52a641c3-5baf-4d1d-a51f-93f372b1013e" (UID: "52a641c3-5baf-4d1d-a51f-93f372b1013e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.143390 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52a641c3-5baf-4d1d-a51f-93f372b1013e-kube-api-access-q2fgr" (OuterVolumeSpecName: "kube-api-access-q2fgr") pod "52a641c3-5baf-4d1d-a51f-93f372b1013e" (UID: "52a641c3-5baf-4d1d-a51f-93f372b1013e"). InnerVolumeSpecName "kube-api-access-q2fgr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.201801 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jk5b2" Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.245298 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q2fgr\" (UniqueName: \"kubernetes.io/projected/52a641c3-5baf-4d1d-a51f-93f372b1013e-kube-api-access-q2fgr\") on node \"crc\" DevicePath \"\"" Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.245331 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52a641c3-5baf-4d1d-a51f-93f372b1013e-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.278036 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/52a641c3-5baf-4d1d-a51f-93f372b1013e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "52a641c3-5baf-4d1d-a51f-93f372b1013e" (UID: "52a641c3-5baf-4d1d-a51f-93f372b1013e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.346134 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-twtp2\" (UniqueName: \"kubernetes.io/projected/4f03c333-e449-4a18-8efb-f8834911a599-kube-api-access-twtp2\") pod \"4f03c333-e449-4a18-8efb-f8834911a599\" (UID: \"4f03c333-e449-4a18-8efb-f8834911a599\") " Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.346227 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f03c333-e449-4a18-8efb-f8834911a599-catalog-content\") pod \"4f03c333-e449-4a18-8efb-f8834911a599\" (UID: \"4f03c333-e449-4a18-8efb-f8834911a599\") " Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.346377 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f03c333-e449-4a18-8efb-f8834911a599-utilities\") pod \"4f03c333-e449-4a18-8efb-f8834911a599\" (UID: \"4f03c333-e449-4a18-8efb-f8834911a599\") " Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.346648 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52a641c3-5baf-4d1d-a51f-93f372b1013e-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.348087 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4f03c333-e449-4a18-8efb-f8834911a599-utilities" (OuterVolumeSpecName: "utilities") pod "4f03c333-e449-4a18-8efb-f8834911a599" (UID: "4f03c333-e449-4a18-8efb-f8834911a599"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.351379 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f03c333-e449-4a18-8efb-f8834911a599-kube-api-access-twtp2" (OuterVolumeSpecName: "kube-api-access-twtp2") pod "4f03c333-e449-4a18-8efb-f8834911a599" (UID: "4f03c333-e449-4a18-8efb-f8834911a599"). InnerVolumeSpecName "kube-api-access-twtp2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.375662 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ps6dr"] Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.401298 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4f03c333-e449-4a18-8efb-f8834911a599-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4f03c333-e449-4a18-8efb-f8834911a599" (UID: "4f03c333-e449-4a18-8efb-f8834911a599"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.451984 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-twtp2\" (UniqueName: \"kubernetes.io/projected/4f03c333-e449-4a18-8efb-f8834911a599-kube-api-access-twtp2\") on node \"crc\" DevicePath \"\"" Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.452028 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f03c333-e449-4a18-8efb-f8834911a599-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.452039 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f03c333-e449-4a18-8efb-f8834911a599-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.811375 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ps6dr" event={"ID":"babbc7b3-6e09-43a7-9cbc-70ba7076ab9f","Type":"ContainerStarted","Data":"f96268ce64401e6de4670c6a7db33813d86778b4883871e7d5ae166813600b19"} Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.811424 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ps6dr" event={"ID":"babbc7b3-6e09-43a7-9cbc-70ba7076ab9f","Type":"ContainerStarted","Data":"e5797eb932a11d670dfe813565302e2de268d2eaf86868c553352724b87d6466"} Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.812347 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-ps6dr" Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.814446 4733 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-ps6dr container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.64:8080/healthz\": dial tcp 10.217.0.64:8080: connect: connection refused" start-of-body= Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.814491 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-ps6dr" podUID="babbc7b3-6e09-43a7-9cbc-70ba7076ab9f" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.64:8080/healthz\": dial tcp 10.217.0.64:8080: connect: connection refused" Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.815197 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v5t9s" event={"ID":"7b6ed8e1-1992-4d83-954a-552aa8e7acb3","Type":"ContainerDied","Data":"6f602392a11e5bfe8ca1c7e652ed4c15847974f2580cb5dfd7dd62af9c7f5dd2"} Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.815251 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v5t9s" Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.815349 4733 scope.go:117] "RemoveContainer" containerID="c9aa09ab01768bd69cd858d98490db679284ae40c82f880a58ece8b87c424d3f" Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.817664 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7gm8t" event={"ID":"d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2","Type":"ContainerDied","Data":"e71d638d6e432213eeabe61563af42a90b1c58fca72b381e6d73aa56388e0832"} Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.817764 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7gm8t" Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.828938 4733 generic.go:334] "Generic (PLEG): container finished" podID="4f03c333-e449-4a18-8efb-f8834911a599" containerID="e4a4c0a9fa82721f255c6e6572b720748b1dd22edba932206f6daaf366db2c53" exitCode=0 Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.829037 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jk5b2" Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.829070 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jk5b2" event={"ID":"4f03c333-e449-4a18-8efb-f8834911a599","Type":"ContainerDied","Data":"e4a4c0a9fa82721f255c6e6572b720748b1dd22edba932206f6daaf366db2c53"} Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.829105 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jk5b2" event={"ID":"4f03c333-e449-4a18-8efb-f8834911a599","Type":"ContainerDied","Data":"751df3ff976e06527b60c7efd19e9458f338ee5a7817ab06b03af72f184f341e"} Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.841953 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-csdnv" event={"ID":"ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e","Type":"ContainerDied","Data":"7e43f999c0810719691f84fda6c80e5d6a519cb83d75ac0f3660b1a7597c4472"} Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.842071 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-ps6dr" podStartSLOduration=1.8420531100000002 podStartE2EDuration="1.84205311s" podCreationTimestamp="2026-02-02 15:21:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:21:20.836866353 +0000 UTC m=+364.288327761" watchObservedRunningTime="2026-02-02 15:21:20.84205311 +0000 UTC m=+364.293514468" Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.842105 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-csdnv" Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.846491 4733 scope.go:117] "RemoveContainer" containerID="b217ce646f132f2a839baaa49cbb8544b12840ec5f9d6e1bf39efc02b87cec16" Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.850435 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9fckn" event={"ID":"52a641c3-5baf-4d1d-a51f-93f372b1013e","Type":"ContainerDied","Data":"000767d15de76401bdc4d99e1b55b5ee2cbb57e6b2c1ffb4177b58524bd5a939"} Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.850594 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9fckn" Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.870205 4733 scope.go:117] "RemoveContainer" containerID="76efd077689c639193102dc831e6ccbaa0f0fdcbeff0a86ff4008a3b266fb8b5" Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.874894 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7gm8t"] Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.878357 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7gm8t"] Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.886464 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-v5t9s"] Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.891461 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-v5t9s"] Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.913426 4733 scope.go:117] "RemoveContainer" containerID="720a4dabd902a96d042261a42019bb9ea3beee89af69a1d942baecce6063882e" Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.937241 4733 scope.go:117] "RemoveContainer" containerID="0cdc71ed22d00abcd62c5b8053ec06dea21eb456bf11c35da7d64cdf192f1f83" Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.940305 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-csdnv"] Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.954238 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-csdnv"] Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.958792 4733 scope.go:117] "RemoveContainer" containerID="7c6ed31ff0747875a20b9c5efbf5918d9c17502ec571eca60f281c6b387e8560" Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.962113 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jk5b2"] Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.965511 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jk5b2"] Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.969436 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9fckn"] Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.972532 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9fckn"] Feb 02 15:21:20 crc kubenswrapper[4733]: I0202 15:21:20.984346 4733 scope.go:117] "RemoveContainer" containerID="e4a4c0a9fa82721f255c6e6572b720748b1dd22edba932206f6daaf366db2c53" Feb 02 15:21:21 crc kubenswrapper[4733]: I0202 15:21:21.002257 4733 scope.go:117] "RemoveContainer" containerID="4ebe46184910bfd2a590af1455b2c9e53b259a81917e5aebabb1c387b30f8898" Feb 02 15:21:21 crc kubenswrapper[4733]: I0202 15:21:21.018686 4733 scope.go:117] "RemoveContainer" containerID="85556a9a2478a4e15cd0900c51c32868a05b1d6954d5ce370f83dba7566ff3e4" Feb 02 15:21:21 crc kubenswrapper[4733]: I0202 15:21:21.041132 4733 scope.go:117] "RemoveContainer" containerID="e4a4c0a9fa82721f255c6e6572b720748b1dd22edba932206f6daaf366db2c53" Feb 02 15:21:21 crc kubenswrapper[4733]: E0202 15:21:21.041596 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4a4c0a9fa82721f255c6e6572b720748b1dd22edba932206f6daaf366db2c53\": container with ID starting with e4a4c0a9fa82721f255c6e6572b720748b1dd22edba932206f6daaf366db2c53 not found: ID does not exist" containerID="e4a4c0a9fa82721f255c6e6572b720748b1dd22edba932206f6daaf366db2c53" Feb 02 15:21:21 crc kubenswrapper[4733]: I0202 15:21:21.041636 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4a4c0a9fa82721f255c6e6572b720748b1dd22edba932206f6daaf366db2c53"} err="failed to get container status \"e4a4c0a9fa82721f255c6e6572b720748b1dd22edba932206f6daaf366db2c53\": rpc error: code = NotFound desc = could not find container \"e4a4c0a9fa82721f255c6e6572b720748b1dd22edba932206f6daaf366db2c53\": container with ID starting with e4a4c0a9fa82721f255c6e6572b720748b1dd22edba932206f6daaf366db2c53 not found: ID does not exist" Feb 02 15:21:21 crc kubenswrapper[4733]: I0202 15:21:21.041662 4733 scope.go:117] "RemoveContainer" containerID="4ebe46184910bfd2a590af1455b2c9e53b259a81917e5aebabb1c387b30f8898" Feb 02 15:21:21 crc kubenswrapper[4733]: E0202 15:21:21.042128 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ebe46184910bfd2a590af1455b2c9e53b259a81917e5aebabb1c387b30f8898\": container with ID starting with 4ebe46184910bfd2a590af1455b2c9e53b259a81917e5aebabb1c387b30f8898 not found: ID does not exist" containerID="4ebe46184910bfd2a590af1455b2c9e53b259a81917e5aebabb1c387b30f8898" Feb 02 15:21:21 crc kubenswrapper[4733]: I0202 15:21:21.042178 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ebe46184910bfd2a590af1455b2c9e53b259a81917e5aebabb1c387b30f8898"} err="failed to get container status \"4ebe46184910bfd2a590af1455b2c9e53b259a81917e5aebabb1c387b30f8898\": rpc error: code = NotFound desc = could not find container \"4ebe46184910bfd2a590af1455b2c9e53b259a81917e5aebabb1c387b30f8898\": container with ID starting with 4ebe46184910bfd2a590af1455b2c9e53b259a81917e5aebabb1c387b30f8898 not found: ID does not exist" Feb 02 15:21:21 crc kubenswrapper[4733]: I0202 15:21:21.042202 4733 scope.go:117] "RemoveContainer" containerID="85556a9a2478a4e15cd0900c51c32868a05b1d6954d5ce370f83dba7566ff3e4" Feb 02 15:21:21 crc kubenswrapper[4733]: E0202 15:21:21.042476 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85556a9a2478a4e15cd0900c51c32868a05b1d6954d5ce370f83dba7566ff3e4\": container with ID starting with 85556a9a2478a4e15cd0900c51c32868a05b1d6954d5ce370f83dba7566ff3e4 not found: ID does not exist" containerID="85556a9a2478a4e15cd0900c51c32868a05b1d6954d5ce370f83dba7566ff3e4" Feb 02 15:21:21 crc kubenswrapper[4733]: I0202 15:21:21.042568 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85556a9a2478a4e15cd0900c51c32868a05b1d6954d5ce370f83dba7566ff3e4"} err="failed to get container status \"85556a9a2478a4e15cd0900c51c32868a05b1d6954d5ce370f83dba7566ff3e4\": rpc error: code = NotFound desc = could not find container \"85556a9a2478a4e15cd0900c51c32868a05b1d6954d5ce370f83dba7566ff3e4\": container with ID starting with 85556a9a2478a4e15cd0900c51c32868a05b1d6954d5ce370f83dba7566ff3e4 not found: ID does not exist" Feb 02 15:21:21 crc kubenswrapper[4733]: I0202 15:21:21.042647 4733 scope.go:117] "RemoveContainer" containerID="35db594c74e3ce27d2532036e1ab3f1815c9fed37332b6cab1836020121772db" Feb 02 15:21:21 crc kubenswrapper[4733]: I0202 15:21:21.055033 4733 scope.go:117] "RemoveContainer" containerID="bc16b500617a150f81ab1c82995e4be4e445957acddb58d55773c44c85ff85e7" Feb 02 15:21:21 crc kubenswrapper[4733]: I0202 15:21:21.067650 4733 scope.go:117] "RemoveContainer" containerID="444ff31d550de20fdebc035f8168dc30ce053ded9043e88036b8f6cb1f3adde4" Feb 02 15:21:21 crc kubenswrapper[4733]: I0202 15:21:21.081048 4733 scope.go:117] "RemoveContainer" containerID="b598793b0ec7ec274a338dccfac6c1b7966f959c1eb4170470db97f776a45622" Feb 02 15:21:21 crc kubenswrapper[4733]: I0202 15:21:21.261771 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f03c333-e449-4a18-8efb-f8834911a599" path="/var/lib/kubelet/pods/4f03c333-e449-4a18-8efb-f8834911a599/volumes" Feb 02 15:21:21 crc kubenswrapper[4733]: I0202 15:21:21.263533 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52a641c3-5baf-4d1d-a51f-93f372b1013e" path="/var/lib/kubelet/pods/52a641c3-5baf-4d1d-a51f-93f372b1013e/volumes" Feb 02 15:21:21 crc kubenswrapper[4733]: I0202 15:21:21.265024 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b6ed8e1-1992-4d83-954a-552aa8e7acb3" path="/var/lib/kubelet/pods/7b6ed8e1-1992-4d83-954a-552aa8e7acb3/volumes" Feb 02 15:21:21 crc kubenswrapper[4733]: I0202 15:21:21.267076 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2" path="/var/lib/kubelet/pods/d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2/volumes" Feb 02 15:21:21 crc kubenswrapper[4733]: I0202 15:21:21.268383 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e" path="/var/lib/kubelet/pods/ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e/volumes" Feb 02 15:21:21 crc kubenswrapper[4733]: I0202 15:21:21.865901 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-ps6dr" Feb 02 15:21:31 crc kubenswrapper[4733]: I0202 15:21:31.521505 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mbh5r"] Feb 02 15:21:31 crc kubenswrapper[4733]: E0202 15:21:31.523038 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f03c333-e449-4a18-8efb-f8834911a599" containerName="extract-utilities" Feb 02 15:21:31 crc kubenswrapper[4733]: I0202 15:21:31.523139 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f03c333-e449-4a18-8efb-f8834911a599" containerName="extract-utilities" Feb 02 15:21:31 crc kubenswrapper[4733]: E0202 15:21:31.523278 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b6ed8e1-1992-4d83-954a-552aa8e7acb3" containerName="registry-server" Feb 02 15:21:31 crc kubenswrapper[4733]: I0202 15:21:31.523354 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b6ed8e1-1992-4d83-954a-552aa8e7acb3" containerName="registry-server" Feb 02 15:21:31 crc kubenswrapper[4733]: E0202 15:21:31.523414 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2" containerName="extract-utilities" Feb 02 15:21:31 crc kubenswrapper[4733]: I0202 15:21:31.523473 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2" containerName="extract-utilities" Feb 02 15:21:31 crc kubenswrapper[4733]: E0202 15:21:31.523536 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2" containerName="registry-server" Feb 02 15:21:31 crc kubenswrapper[4733]: I0202 15:21:31.523597 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2" containerName="registry-server" Feb 02 15:21:31 crc kubenswrapper[4733]: E0202 15:21:31.523657 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f03c333-e449-4a18-8efb-f8834911a599" containerName="extract-content" Feb 02 15:21:31 crc kubenswrapper[4733]: I0202 15:21:31.523708 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f03c333-e449-4a18-8efb-f8834911a599" containerName="extract-content" Feb 02 15:21:31 crc kubenswrapper[4733]: E0202 15:21:31.523768 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2" containerName="extract-content" Feb 02 15:21:31 crc kubenswrapper[4733]: I0202 15:21:31.523826 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2" containerName="extract-content" Feb 02 15:21:31 crc kubenswrapper[4733]: E0202 15:21:31.523886 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e" containerName="marketplace-operator" Feb 02 15:21:31 crc kubenswrapper[4733]: I0202 15:21:31.523940 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e" containerName="marketplace-operator" Feb 02 15:21:31 crc kubenswrapper[4733]: E0202 15:21:31.523999 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52a641c3-5baf-4d1d-a51f-93f372b1013e" containerName="registry-server" Feb 02 15:21:31 crc kubenswrapper[4733]: I0202 15:21:31.524053 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="52a641c3-5baf-4d1d-a51f-93f372b1013e" containerName="registry-server" Feb 02 15:21:31 crc kubenswrapper[4733]: E0202 15:21:31.524113 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b6ed8e1-1992-4d83-954a-552aa8e7acb3" containerName="extract-utilities" Feb 02 15:21:31 crc kubenswrapper[4733]: I0202 15:21:31.524190 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b6ed8e1-1992-4d83-954a-552aa8e7acb3" containerName="extract-utilities" Feb 02 15:21:31 crc kubenswrapper[4733]: E0202 15:21:31.524263 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f03c333-e449-4a18-8efb-f8834911a599" containerName="registry-server" Feb 02 15:21:31 crc kubenswrapper[4733]: I0202 15:21:31.524322 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f03c333-e449-4a18-8efb-f8834911a599" containerName="registry-server" Feb 02 15:21:31 crc kubenswrapper[4733]: E0202 15:21:31.524410 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52a641c3-5baf-4d1d-a51f-93f372b1013e" containerName="extract-utilities" Feb 02 15:21:31 crc kubenswrapper[4733]: I0202 15:21:31.524478 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="52a641c3-5baf-4d1d-a51f-93f372b1013e" containerName="extract-utilities" Feb 02 15:21:31 crc kubenswrapper[4733]: E0202 15:21:31.524537 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b6ed8e1-1992-4d83-954a-552aa8e7acb3" containerName="extract-content" Feb 02 15:21:31 crc kubenswrapper[4733]: I0202 15:21:31.524589 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b6ed8e1-1992-4d83-954a-552aa8e7acb3" containerName="extract-content" Feb 02 15:21:31 crc kubenswrapper[4733]: E0202 15:21:31.524653 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52a641c3-5baf-4d1d-a51f-93f372b1013e" containerName="extract-content" Feb 02 15:21:31 crc kubenswrapper[4733]: I0202 15:21:31.524706 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="52a641c3-5baf-4d1d-a51f-93f372b1013e" containerName="extract-content" Feb 02 15:21:31 crc kubenswrapper[4733]: I0202 15:21:31.524839 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="52a641c3-5baf-4d1d-a51f-93f372b1013e" containerName="registry-server" Feb 02 15:21:31 crc kubenswrapper[4733]: I0202 15:21:31.524902 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f03c333-e449-4a18-8efb-f8834911a599" containerName="registry-server" Feb 02 15:21:31 crc kubenswrapper[4733]: I0202 15:21:31.524958 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b6ed8e1-1992-4d83-954a-552aa8e7acb3" containerName="registry-server" Feb 02 15:21:31 crc kubenswrapper[4733]: I0202 15:21:31.525023 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea1f4b1c-5a8a-42ac-b1d7-784cbb95012e" containerName="marketplace-operator" Feb 02 15:21:31 crc kubenswrapper[4733]: I0202 15:21:31.525079 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9d9f8ab-a82c-4994-a1fd-b62d73eb48d2" containerName="registry-server" Feb 02 15:21:31 crc kubenswrapper[4733]: I0202 15:21:31.525776 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mbh5r" Feb 02 15:21:31 crc kubenswrapper[4733]: I0202 15:21:31.527979 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 02 15:21:31 crc kubenswrapper[4733]: I0202 15:21:31.543903 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mbh5r"] Feb 02 15:21:31 crc kubenswrapper[4733]: I0202 15:21:31.597632 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvvwn\" (UniqueName: \"kubernetes.io/projected/b8fd49b0-ac5d-4b93-8c03-7ac35c150aa2-kube-api-access-hvvwn\") pod \"certified-operators-mbh5r\" (UID: \"b8fd49b0-ac5d-4b93-8c03-7ac35c150aa2\") " pod="openshift-marketplace/certified-operators-mbh5r" Feb 02 15:21:31 crc kubenswrapper[4733]: I0202 15:21:31.597713 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8fd49b0-ac5d-4b93-8c03-7ac35c150aa2-catalog-content\") pod \"certified-operators-mbh5r\" (UID: \"b8fd49b0-ac5d-4b93-8c03-7ac35c150aa2\") " pod="openshift-marketplace/certified-operators-mbh5r" Feb 02 15:21:31 crc kubenswrapper[4733]: I0202 15:21:31.597739 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8fd49b0-ac5d-4b93-8c03-7ac35c150aa2-utilities\") pod \"certified-operators-mbh5r\" (UID: \"b8fd49b0-ac5d-4b93-8c03-7ac35c150aa2\") " pod="openshift-marketplace/certified-operators-mbh5r" Feb 02 15:21:31 crc kubenswrapper[4733]: I0202 15:21:31.698889 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8fd49b0-ac5d-4b93-8c03-7ac35c150aa2-catalog-content\") pod \"certified-operators-mbh5r\" (UID: \"b8fd49b0-ac5d-4b93-8c03-7ac35c150aa2\") " pod="openshift-marketplace/certified-operators-mbh5r" Feb 02 15:21:31 crc kubenswrapper[4733]: I0202 15:21:31.699410 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8fd49b0-ac5d-4b93-8c03-7ac35c150aa2-utilities\") pod \"certified-operators-mbh5r\" (UID: \"b8fd49b0-ac5d-4b93-8c03-7ac35c150aa2\") " pod="openshift-marketplace/certified-operators-mbh5r" Feb 02 15:21:31 crc kubenswrapper[4733]: I0202 15:21:31.699546 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvvwn\" (UniqueName: \"kubernetes.io/projected/b8fd49b0-ac5d-4b93-8c03-7ac35c150aa2-kube-api-access-hvvwn\") pod \"certified-operators-mbh5r\" (UID: \"b8fd49b0-ac5d-4b93-8c03-7ac35c150aa2\") " pod="openshift-marketplace/certified-operators-mbh5r" Feb 02 15:21:31 crc kubenswrapper[4733]: I0202 15:21:31.699848 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8fd49b0-ac5d-4b93-8c03-7ac35c150aa2-catalog-content\") pod \"certified-operators-mbh5r\" (UID: \"b8fd49b0-ac5d-4b93-8c03-7ac35c150aa2\") " pod="openshift-marketplace/certified-operators-mbh5r" Feb 02 15:21:31 crc kubenswrapper[4733]: I0202 15:21:31.700023 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8fd49b0-ac5d-4b93-8c03-7ac35c150aa2-utilities\") pod \"certified-operators-mbh5r\" (UID: \"b8fd49b0-ac5d-4b93-8c03-7ac35c150aa2\") " pod="openshift-marketplace/certified-operators-mbh5r" Feb 02 15:21:31 crc kubenswrapper[4733]: I0202 15:21:31.737975 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvvwn\" (UniqueName: \"kubernetes.io/projected/b8fd49b0-ac5d-4b93-8c03-7ac35c150aa2-kube-api-access-hvvwn\") pod \"certified-operators-mbh5r\" (UID: \"b8fd49b0-ac5d-4b93-8c03-7ac35c150aa2\") " pod="openshift-marketplace/certified-operators-mbh5r" Feb 02 15:21:31 crc kubenswrapper[4733]: I0202 15:21:31.751554 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-qvwpw"] Feb 02 15:21:31 crc kubenswrapper[4733]: I0202 15:21:31.753779 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qvwpw" Feb 02 15:21:31 crc kubenswrapper[4733]: I0202 15:21:31.755293 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 02 15:21:31 crc kubenswrapper[4733]: I0202 15:21:31.762405 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qvwpw"] Feb 02 15:21:31 crc kubenswrapper[4733]: I0202 15:21:31.800523 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54e05a7f-a312-4455-83ec-be97bf446b68-utilities\") pod \"community-operators-qvwpw\" (UID: \"54e05a7f-a312-4455-83ec-be97bf446b68\") " pod="openshift-marketplace/community-operators-qvwpw" Feb 02 15:21:31 crc kubenswrapper[4733]: I0202 15:21:31.800561 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54e05a7f-a312-4455-83ec-be97bf446b68-catalog-content\") pod \"community-operators-qvwpw\" (UID: \"54e05a7f-a312-4455-83ec-be97bf446b68\") " pod="openshift-marketplace/community-operators-qvwpw" Feb 02 15:21:31 crc kubenswrapper[4733]: I0202 15:21:31.800658 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sg6p2\" (UniqueName: \"kubernetes.io/projected/54e05a7f-a312-4455-83ec-be97bf446b68-kube-api-access-sg6p2\") pod \"community-operators-qvwpw\" (UID: \"54e05a7f-a312-4455-83ec-be97bf446b68\") " pod="openshift-marketplace/community-operators-qvwpw" Feb 02 15:21:31 crc kubenswrapper[4733]: I0202 15:21:31.859026 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mbh5r" Feb 02 15:21:31 crc kubenswrapper[4733]: I0202 15:21:31.901899 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54e05a7f-a312-4455-83ec-be97bf446b68-utilities\") pod \"community-operators-qvwpw\" (UID: \"54e05a7f-a312-4455-83ec-be97bf446b68\") " pod="openshift-marketplace/community-operators-qvwpw" Feb 02 15:21:31 crc kubenswrapper[4733]: I0202 15:21:31.902088 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54e05a7f-a312-4455-83ec-be97bf446b68-catalog-content\") pod \"community-operators-qvwpw\" (UID: \"54e05a7f-a312-4455-83ec-be97bf446b68\") " pod="openshift-marketplace/community-operators-qvwpw" Feb 02 15:21:31 crc kubenswrapper[4733]: I0202 15:21:31.902230 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sg6p2\" (UniqueName: \"kubernetes.io/projected/54e05a7f-a312-4455-83ec-be97bf446b68-kube-api-access-sg6p2\") pod \"community-operators-qvwpw\" (UID: \"54e05a7f-a312-4455-83ec-be97bf446b68\") " pod="openshift-marketplace/community-operators-qvwpw" Feb 02 15:21:31 crc kubenswrapper[4733]: I0202 15:21:31.902778 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54e05a7f-a312-4455-83ec-be97bf446b68-utilities\") pod \"community-operators-qvwpw\" (UID: \"54e05a7f-a312-4455-83ec-be97bf446b68\") " pod="openshift-marketplace/community-operators-qvwpw" Feb 02 15:21:31 crc kubenswrapper[4733]: I0202 15:21:31.902997 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54e05a7f-a312-4455-83ec-be97bf446b68-catalog-content\") pod \"community-operators-qvwpw\" (UID: \"54e05a7f-a312-4455-83ec-be97bf446b68\") " pod="openshift-marketplace/community-operators-qvwpw" Feb 02 15:21:31 crc kubenswrapper[4733]: I0202 15:21:31.927962 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sg6p2\" (UniqueName: \"kubernetes.io/projected/54e05a7f-a312-4455-83ec-be97bf446b68-kube-api-access-sg6p2\") pod \"community-operators-qvwpw\" (UID: \"54e05a7f-a312-4455-83ec-be97bf446b68\") " pod="openshift-marketplace/community-operators-qvwpw" Feb 02 15:21:32 crc kubenswrapper[4733]: I0202 15:21:32.076661 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qvwpw" Feb 02 15:21:32 crc kubenswrapper[4733]: I0202 15:21:32.261084 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mbh5r"] Feb 02 15:21:32 crc kubenswrapper[4733]: I0202 15:21:32.285815 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qvwpw"] Feb 02 15:21:32 crc kubenswrapper[4733]: W0202 15:21:32.305758 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod54e05a7f_a312_4455_83ec_be97bf446b68.slice/crio-52a2e2e20f937d5f3f36c9af400c488c773b201b0ead2bc9b5b7b281290479d2 WatchSource:0}: Error finding container 52a2e2e20f937d5f3f36c9af400c488c773b201b0ead2bc9b5b7b281290479d2: Status 404 returned error can't find the container with id 52a2e2e20f937d5f3f36c9af400c488c773b201b0ead2bc9b5b7b281290479d2 Feb 02 15:21:32 crc kubenswrapper[4733]: I0202 15:21:32.401488 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-54464559b6-tpnl9"] Feb 02 15:21:32 crc kubenswrapper[4733]: I0202 15:21:32.401683 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-54464559b6-tpnl9" podUID="e1407e69-2431-4fdf-b8a7-b6396b46b6d9" containerName="controller-manager" containerID="cri-o://6747c0d370cf58da2862f3779f20a77bc346d0aaaf4784f3e5c73432169736b7" gracePeriod=30 Feb 02 15:21:32 crc kubenswrapper[4733]: I0202 15:21:32.729109 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-54464559b6-tpnl9" Feb 02 15:21:32 crc kubenswrapper[4733]: I0202 15:21:32.811918 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e1407e69-2431-4fdf-b8a7-b6396b46b6d9-client-ca\") pod \"e1407e69-2431-4fdf-b8a7-b6396b46b6d9\" (UID: \"e1407e69-2431-4fdf-b8a7-b6396b46b6d9\") " Feb 02 15:21:32 crc kubenswrapper[4733]: I0202 15:21:32.811959 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1407e69-2431-4fdf-b8a7-b6396b46b6d9-config\") pod \"e1407e69-2431-4fdf-b8a7-b6396b46b6d9\" (UID: \"e1407e69-2431-4fdf-b8a7-b6396b46b6d9\") " Feb 02 15:21:32 crc kubenswrapper[4733]: I0202 15:21:32.812042 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e1407e69-2431-4fdf-b8a7-b6396b46b6d9-proxy-ca-bundles\") pod \"e1407e69-2431-4fdf-b8a7-b6396b46b6d9\" (UID: \"e1407e69-2431-4fdf-b8a7-b6396b46b6d9\") " Feb 02 15:21:32 crc kubenswrapper[4733]: I0202 15:21:32.812073 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e1407e69-2431-4fdf-b8a7-b6396b46b6d9-serving-cert\") pod \"e1407e69-2431-4fdf-b8a7-b6396b46b6d9\" (UID: \"e1407e69-2431-4fdf-b8a7-b6396b46b6d9\") " Feb 02 15:21:32 crc kubenswrapper[4733]: I0202 15:21:32.812102 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nj2s4\" (UniqueName: \"kubernetes.io/projected/e1407e69-2431-4fdf-b8a7-b6396b46b6d9-kube-api-access-nj2s4\") pod \"e1407e69-2431-4fdf-b8a7-b6396b46b6d9\" (UID: \"e1407e69-2431-4fdf-b8a7-b6396b46b6d9\") " Feb 02 15:21:32 crc kubenswrapper[4733]: I0202 15:21:32.812801 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1407e69-2431-4fdf-b8a7-b6396b46b6d9-client-ca" (OuterVolumeSpecName: "client-ca") pod "e1407e69-2431-4fdf-b8a7-b6396b46b6d9" (UID: "e1407e69-2431-4fdf-b8a7-b6396b46b6d9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:21:32 crc kubenswrapper[4733]: I0202 15:21:32.812825 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1407e69-2431-4fdf-b8a7-b6396b46b6d9-config" (OuterVolumeSpecName: "config") pod "e1407e69-2431-4fdf-b8a7-b6396b46b6d9" (UID: "e1407e69-2431-4fdf-b8a7-b6396b46b6d9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:21:32 crc kubenswrapper[4733]: I0202 15:21:32.812810 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1407e69-2431-4fdf-b8a7-b6396b46b6d9-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "e1407e69-2431-4fdf-b8a7-b6396b46b6d9" (UID: "e1407e69-2431-4fdf-b8a7-b6396b46b6d9"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:21:32 crc kubenswrapper[4733]: I0202 15:21:32.817150 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1407e69-2431-4fdf-b8a7-b6396b46b6d9-kube-api-access-nj2s4" (OuterVolumeSpecName: "kube-api-access-nj2s4") pod "e1407e69-2431-4fdf-b8a7-b6396b46b6d9" (UID: "e1407e69-2431-4fdf-b8a7-b6396b46b6d9"). InnerVolumeSpecName "kube-api-access-nj2s4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:21:32 crc kubenswrapper[4733]: I0202 15:21:32.821569 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1407e69-2431-4fdf-b8a7-b6396b46b6d9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e1407e69-2431-4fdf-b8a7-b6396b46b6d9" (UID: "e1407e69-2431-4fdf-b8a7-b6396b46b6d9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:21:32 crc kubenswrapper[4733]: I0202 15:21:32.913229 4733 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e1407e69-2431-4fdf-b8a7-b6396b46b6d9-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 02 15:21:32 crc kubenswrapper[4733]: I0202 15:21:32.913261 4733 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e1407e69-2431-4fdf-b8a7-b6396b46b6d9-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 02 15:21:32 crc kubenswrapper[4733]: I0202 15:21:32.913272 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nj2s4\" (UniqueName: \"kubernetes.io/projected/e1407e69-2431-4fdf-b8a7-b6396b46b6d9-kube-api-access-nj2s4\") on node \"crc\" DevicePath \"\"" Feb 02 15:21:32 crc kubenswrapper[4733]: I0202 15:21:32.913284 4733 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1407e69-2431-4fdf-b8a7-b6396b46b6d9-config\") on node \"crc\" DevicePath \"\"" Feb 02 15:21:32 crc kubenswrapper[4733]: I0202 15:21:32.913293 4733 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e1407e69-2431-4fdf-b8a7-b6396b46b6d9-client-ca\") on node \"crc\" DevicePath \"\"" Feb 02 15:21:32 crc kubenswrapper[4733]: I0202 15:21:32.967395 4733 generic.go:334] "Generic (PLEG): container finished" podID="e1407e69-2431-4fdf-b8a7-b6396b46b6d9" containerID="6747c0d370cf58da2862f3779f20a77bc346d0aaaf4784f3e5c73432169736b7" exitCode=0 Feb 02 15:21:32 crc kubenswrapper[4733]: I0202 15:21:32.967460 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-54464559b6-tpnl9" event={"ID":"e1407e69-2431-4fdf-b8a7-b6396b46b6d9","Type":"ContainerDied","Data":"6747c0d370cf58da2862f3779f20a77bc346d0aaaf4784f3e5c73432169736b7"} Feb 02 15:21:32 crc kubenswrapper[4733]: I0202 15:21:32.967491 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-54464559b6-tpnl9" Feb 02 15:21:32 crc kubenswrapper[4733]: I0202 15:21:32.967509 4733 scope.go:117] "RemoveContainer" containerID="6747c0d370cf58da2862f3779f20a77bc346d0aaaf4784f3e5c73432169736b7" Feb 02 15:21:32 crc kubenswrapper[4733]: I0202 15:21:32.967498 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-54464559b6-tpnl9" event={"ID":"e1407e69-2431-4fdf-b8a7-b6396b46b6d9","Type":"ContainerDied","Data":"56ad4feb0f6070978a7683f08cb2b131edb0adc8841dc42500ac1796c06f6d4f"} Feb 02 15:21:32 crc kubenswrapper[4733]: I0202 15:21:32.969868 4733 generic.go:334] "Generic (PLEG): container finished" podID="54e05a7f-a312-4455-83ec-be97bf446b68" containerID="cab3c6a927000cc92e9be8f28e0f6fcbefd2185e90b8726367bf7125d691a901" exitCode=0 Feb 02 15:21:32 crc kubenswrapper[4733]: I0202 15:21:32.969946 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qvwpw" event={"ID":"54e05a7f-a312-4455-83ec-be97bf446b68","Type":"ContainerDied","Data":"cab3c6a927000cc92e9be8f28e0f6fcbefd2185e90b8726367bf7125d691a901"} Feb 02 15:21:32 crc kubenswrapper[4733]: I0202 15:21:32.969972 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qvwpw" event={"ID":"54e05a7f-a312-4455-83ec-be97bf446b68","Type":"ContainerStarted","Data":"52a2e2e20f937d5f3f36c9af400c488c773b201b0ead2bc9b5b7b281290479d2"} Feb 02 15:21:32 crc kubenswrapper[4733]: I0202 15:21:32.978832 4733 generic.go:334] "Generic (PLEG): container finished" podID="b8fd49b0-ac5d-4b93-8c03-7ac35c150aa2" containerID="1cbdd32e3127688eb085ebe492b7e5a2bc5319d7fce8b1c78f55f10aa24296f5" exitCode=0 Feb 02 15:21:32 crc kubenswrapper[4733]: I0202 15:21:32.978847 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mbh5r" event={"ID":"b8fd49b0-ac5d-4b93-8c03-7ac35c150aa2","Type":"ContainerDied","Data":"1cbdd32e3127688eb085ebe492b7e5a2bc5319d7fce8b1c78f55f10aa24296f5"} Feb 02 15:21:32 crc kubenswrapper[4733]: I0202 15:21:32.978909 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mbh5r" event={"ID":"b8fd49b0-ac5d-4b93-8c03-7ac35c150aa2","Type":"ContainerStarted","Data":"5b5f201bfca263d687aa651c888b3626a6497ad7c468391c4bfecda059e498fc"} Feb 02 15:21:32 crc kubenswrapper[4733]: I0202 15:21:32.990325 4733 scope.go:117] "RemoveContainer" containerID="6747c0d370cf58da2862f3779f20a77bc346d0aaaf4784f3e5c73432169736b7" Feb 02 15:21:32 crc kubenswrapper[4733]: E0202 15:21:32.992224 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6747c0d370cf58da2862f3779f20a77bc346d0aaaf4784f3e5c73432169736b7\": container with ID starting with 6747c0d370cf58da2862f3779f20a77bc346d0aaaf4784f3e5c73432169736b7 not found: ID does not exist" containerID="6747c0d370cf58da2862f3779f20a77bc346d0aaaf4784f3e5c73432169736b7" Feb 02 15:21:32 crc kubenswrapper[4733]: I0202 15:21:32.992263 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6747c0d370cf58da2862f3779f20a77bc346d0aaaf4784f3e5c73432169736b7"} err="failed to get container status \"6747c0d370cf58da2862f3779f20a77bc346d0aaaf4784f3e5c73432169736b7\": rpc error: code = NotFound desc = could not find container \"6747c0d370cf58da2862f3779f20a77bc346d0aaaf4784f3e5c73432169736b7\": container with ID starting with 6747c0d370cf58da2862f3779f20a77bc346d0aaaf4784f3e5c73432169736b7 not found: ID does not exist" Feb 02 15:21:33 crc kubenswrapper[4733]: I0202 15:21:33.028139 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-54464559b6-tpnl9"] Feb 02 15:21:33 crc kubenswrapper[4733]: I0202 15:21:33.034812 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-54464559b6-tpnl9"] Feb 02 15:21:33 crc kubenswrapper[4733]: I0202 15:21:33.262809 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1407e69-2431-4fdf-b8a7-b6396b46b6d9" path="/var/lib/kubelet/pods/e1407e69-2431-4fdf-b8a7-b6396b46b6d9/volumes" Feb 02 15:21:33 crc kubenswrapper[4733]: I0202 15:21:33.706684 4733 patch_prober.go:28] interesting pod/controller-manager-54464559b6-tpnl9 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.61:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 02 15:21:33 crc kubenswrapper[4733]: I0202 15:21:33.706955 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-54464559b6-tpnl9" podUID="e1407e69-2431-4fdf-b8a7-b6396b46b6d9" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.61:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 02 15:21:33 crc kubenswrapper[4733]: I0202 15:21:33.845037 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-75dcc9d6d5-dlww8"] Feb 02 15:21:33 crc kubenswrapper[4733]: E0202 15:21:33.845301 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1407e69-2431-4fdf-b8a7-b6396b46b6d9" containerName="controller-manager" Feb 02 15:21:33 crc kubenswrapper[4733]: I0202 15:21:33.845318 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1407e69-2431-4fdf-b8a7-b6396b46b6d9" containerName="controller-manager" Feb 02 15:21:33 crc kubenswrapper[4733]: I0202 15:21:33.845435 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1407e69-2431-4fdf-b8a7-b6396b46b6d9" containerName="controller-manager" Feb 02 15:21:33 crc kubenswrapper[4733]: I0202 15:21:33.846062 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-75dcc9d6d5-dlww8" Feb 02 15:21:33 crc kubenswrapper[4733]: I0202 15:21:33.848898 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 02 15:21:33 crc kubenswrapper[4733]: I0202 15:21:33.849087 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 02 15:21:33 crc kubenswrapper[4733]: I0202 15:21:33.849284 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 02 15:21:33 crc kubenswrapper[4733]: I0202 15:21:33.850012 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 02 15:21:33 crc kubenswrapper[4733]: I0202 15:21:33.851267 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 02 15:21:33 crc kubenswrapper[4733]: I0202 15:21:33.853444 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 02 15:21:33 crc kubenswrapper[4733]: I0202 15:21:33.860782 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-75dcc9d6d5-dlww8"] Feb 02 15:21:33 crc kubenswrapper[4733]: I0202 15:21:33.862962 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 02 15:21:33 crc kubenswrapper[4733]: I0202 15:21:33.924691 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-smvdc"] Feb 02 15:21:33 crc kubenswrapper[4733]: I0202 15:21:33.926128 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-smvdc" Feb 02 15:21:33 crc kubenswrapper[4733]: I0202 15:21:33.929120 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 02 15:21:33 crc kubenswrapper[4733]: I0202 15:21:33.940312 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-smvdc"] Feb 02 15:21:33 crc kubenswrapper[4733]: I0202 15:21:33.948891 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1b138594-1c90-48e0-83a4-e80f3608f0d2-client-ca\") pod \"controller-manager-75dcc9d6d5-dlww8\" (UID: \"1b138594-1c90-48e0-83a4-e80f3608f0d2\") " pod="openshift-controller-manager/controller-manager-75dcc9d6d5-dlww8" Feb 02 15:21:33 crc kubenswrapper[4733]: I0202 15:21:33.948993 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfch2\" (UniqueName: \"kubernetes.io/projected/1b138594-1c90-48e0-83a4-e80f3608f0d2-kube-api-access-bfch2\") pod \"controller-manager-75dcc9d6d5-dlww8\" (UID: \"1b138594-1c90-48e0-83a4-e80f3608f0d2\") " pod="openshift-controller-manager/controller-manager-75dcc9d6d5-dlww8" Feb 02 15:21:33 crc kubenswrapper[4733]: I0202 15:21:33.949086 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1b138594-1c90-48e0-83a4-e80f3608f0d2-proxy-ca-bundles\") pod \"controller-manager-75dcc9d6d5-dlww8\" (UID: \"1b138594-1c90-48e0-83a4-e80f3608f0d2\") " pod="openshift-controller-manager/controller-manager-75dcc9d6d5-dlww8" Feb 02 15:21:33 crc kubenswrapper[4733]: I0202 15:21:33.949130 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1b138594-1c90-48e0-83a4-e80f3608f0d2-serving-cert\") pod \"controller-manager-75dcc9d6d5-dlww8\" (UID: \"1b138594-1c90-48e0-83a4-e80f3608f0d2\") " pod="openshift-controller-manager/controller-manager-75dcc9d6d5-dlww8" Feb 02 15:21:33 crc kubenswrapper[4733]: I0202 15:21:33.949204 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b138594-1c90-48e0-83a4-e80f3608f0d2-config\") pod \"controller-manager-75dcc9d6d5-dlww8\" (UID: \"1b138594-1c90-48e0-83a4-e80f3608f0d2\") " pod="openshift-controller-manager/controller-manager-75dcc9d6d5-dlww8" Feb 02 15:21:33 crc kubenswrapper[4733]: I0202 15:21:33.986520 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qvwpw" event={"ID":"54e05a7f-a312-4455-83ec-be97bf446b68","Type":"ContainerStarted","Data":"d0259ef30d7e89b3164987c2cfe9dc5bd29abba98ab18ad6caca4259d42b943f"} Feb 02 15:21:34 crc kubenswrapper[4733]: I0202 15:21:34.050367 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1b138594-1c90-48e0-83a4-e80f3608f0d2-client-ca\") pod \"controller-manager-75dcc9d6d5-dlww8\" (UID: \"1b138594-1c90-48e0-83a4-e80f3608f0d2\") " pod="openshift-controller-manager/controller-manager-75dcc9d6d5-dlww8" Feb 02 15:21:34 crc kubenswrapper[4733]: I0202 15:21:34.050424 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5mll\" (UniqueName: \"kubernetes.io/projected/0367d1cb-4122-4c7e-b203-d5c4216c6ef5-kube-api-access-l5mll\") pod \"redhat-marketplace-smvdc\" (UID: \"0367d1cb-4122-4c7e-b203-d5c4216c6ef5\") " pod="openshift-marketplace/redhat-marketplace-smvdc" Feb 02 15:21:34 crc kubenswrapper[4733]: I0202 15:21:34.050447 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0367d1cb-4122-4c7e-b203-d5c4216c6ef5-utilities\") pod \"redhat-marketplace-smvdc\" (UID: \"0367d1cb-4122-4c7e-b203-d5c4216c6ef5\") " pod="openshift-marketplace/redhat-marketplace-smvdc" Feb 02 15:21:34 crc kubenswrapper[4733]: I0202 15:21:34.050467 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfch2\" (UniqueName: \"kubernetes.io/projected/1b138594-1c90-48e0-83a4-e80f3608f0d2-kube-api-access-bfch2\") pod \"controller-manager-75dcc9d6d5-dlww8\" (UID: \"1b138594-1c90-48e0-83a4-e80f3608f0d2\") " pod="openshift-controller-manager/controller-manager-75dcc9d6d5-dlww8" Feb 02 15:21:34 crc kubenswrapper[4733]: I0202 15:21:34.050503 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1b138594-1c90-48e0-83a4-e80f3608f0d2-proxy-ca-bundles\") pod \"controller-manager-75dcc9d6d5-dlww8\" (UID: \"1b138594-1c90-48e0-83a4-e80f3608f0d2\") " pod="openshift-controller-manager/controller-manager-75dcc9d6d5-dlww8" Feb 02 15:21:34 crc kubenswrapper[4733]: I0202 15:21:34.050518 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0367d1cb-4122-4c7e-b203-d5c4216c6ef5-catalog-content\") pod \"redhat-marketplace-smvdc\" (UID: \"0367d1cb-4122-4c7e-b203-d5c4216c6ef5\") " pod="openshift-marketplace/redhat-marketplace-smvdc" Feb 02 15:21:34 crc kubenswrapper[4733]: I0202 15:21:34.050536 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1b138594-1c90-48e0-83a4-e80f3608f0d2-serving-cert\") pod \"controller-manager-75dcc9d6d5-dlww8\" (UID: \"1b138594-1c90-48e0-83a4-e80f3608f0d2\") " pod="openshift-controller-manager/controller-manager-75dcc9d6d5-dlww8" Feb 02 15:21:34 crc kubenswrapper[4733]: I0202 15:21:34.050558 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b138594-1c90-48e0-83a4-e80f3608f0d2-config\") pod \"controller-manager-75dcc9d6d5-dlww8\" (UID: \"1b138594-1c90-48e0-83a4-e80f3608f0d2\") " pod="openshift-controller-manager/controller-manager-75dcc9d6d5-dlww8" Feb 02 15:21:34 crc kubenswrapper[4733]: I0202 15:21:34.051190 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1b138594-1c90-48e0-83a4-e80f3608f0d2-client-ca\") pod \"controller-manager-75dcc9d6d5-dlww8\" (UID: \"1b138594-1c90-48e0-83a4-e80f3608f0d2\") " pod="openshift-controller-manager/controller-manager-75dcc9d6d5-dlww8" Feb 02 15:21:34 crc kubenswrapper[4733]: I0202 15:21:34.051837 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b138594-1c90-48e0-83a4-e80f3608f0d2-config\") pod \"controller-manager-75dcc9d6d5-dlww8\" (UID: \"1b138594-1c90-48e0-83a4-e80f3608f0d2\") " pod="openshift-controller-manager/controller-manager-75dcc9d6d5-dlww8" Feb 02 15:21:34 crc kubenswrapper[4733]: I0202 15:21:34.051991 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1b138594-1c90-48e0-83a4-e80f3608f0d2-proxy-ca-bundles\") pod \"controller-manager-75dcc9d6d5-dlww8\" (UID: \"1b138594-1c90-48e0-83a4-e80f3608f0d2\") " pod="openshift-controller-manager/controller-manager-75dcc9d6d5-dlww8" Feb 02 15:21:34 crc kubenswrapper[4733]: I0202 15:21:34.056795 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1b138594-1c90-48e0-83a4-e80f3608f0d2-serving-cert\") pod \"controller-manager-75dcc9d6d5-dlww8\" (UID: \"1b138594-1c90-48e0-83a4-e80f3608f0d2\") " pod="openshift-controller-manager/controller-manager-75dcc9d6d5-dlww8" Feb 02 15:21:34 crc kubenswrapper[4733]: I0202 15:21:34.066239 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfch2\" (UniqueName: \"kubernetes.io/projected/1b138594-1c90-48e0-83a4-e80f3608f0d2-kube-api-access-bfch2\") pod \"controller-manager-75dcc9d6d5-dlww8\" (UID: \"1b138594-1c90-48e0-83a4-e80f3608f0d2\") " pod="openshift-controller-manager/controller-manager-75dcc9d6d5-dlww8" Feb 02 15:21:34 crc kubenswrapper[4733]: I0202 15:21:34.120037 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kqbhk"] Feb 02 15:21:34 crc kubenswrapper[4733]: I0202 15:21:34.121217 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kqbhk" Feb 02 15:21:34 crc kubenswrapper[4733]: I0202 15:21:34.123288 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 02 15:21:34 crc kubenswrapper[4733]: I0202 15:21:34.134079 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kqbhk"] Feb 02 15:21:34 crc kubenswrapper[4733]: I0202 15:21:34.151207 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0367d1cb-4122-4c7e-b203-d5c4216c6ef5-catalog-content\") pod \"redhat-marketplace-smvdc\" (UID: \"0367d1cb-4122-4c7e-b203-d5c4216c6ef5\") " pod="openshift-marketplace/redhat-marketplace-smvdc" Feb 02 15:21:34 crc kubenswrapper[4733]: I0202 15:21:34.151290 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5mll\" (UniqueName: \"kubernetes.io/projected/0367d1cb-4122-4c7e-b203-d5c4216c6ef5-kube-api-access-l5mll\") pod \"redhat-marketplace-smvdc\" (UID: \"0367d1cb-4122-4c7e-b203-d5c4216c6ef5\") " pod="openshift-marketplace/redhat-marketplace-smvdc" Feb 02 15:21:34 crc kubenswrapper[4733]: I0202 15:21:34.151309 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0367d1cb-4122-4c7e-b203-d5c4216c6ef5-utilities\") pod \"redhat-marketplace-smvdc\" (UID: \"0367d1cb-4122-4c7e-b203-d5c4216c6ef5\") " pod="openshift-marketplace/redhat-marketplace-smvdc" Feb 02 15:21:34 crc kubenswrapper[4733]: I0202 15:21:34.151720 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0367d1cb-4122-4c7e-b203-d5c4216c6ef5-catalog-content\") pod \"redhat-marketplace-smvdc\" (UID: \"0367d1cb-4122-4c7e-b203-d5c4216c6ef5\") " pod="openshift-marketplace/redhat-marketplace-smvdc" Feb 02 15:21:34 crc kubenswrapper[4733]: I0202 15:21:34.151767 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0367d1cb-4122-4c7e-b203-d5c4216c6ef5-utilities\") pod \"redhat-marketplace-smvdc\" (UID: \"0367d1cb-4122-4c7e-b203-d5c4216c6ef5\") " pod="openshift-marketplace/redhat-marketplace-smvdc" Feb 02 15:21:34 crc kubenswrapper[4733]: I0202 15:21:34.168989 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5mll\" (UniqueName: \"kubernetes.io/projected/0367d1cb-4122-4c7e-b203-d5c4216c6ef5-kube-api-access-l5mll\") pod \"redhat-marketplace-smvdc\" (UID: \"0367d1cb-4122-4c7e-b203-d5c4216c6ef5\") " pod="openshift-marketplace/redhat-marketplace-smvdc" Feb 02 15:21:34 crc kubenswrapper[4733]: I0202 15:21:34.171740 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-75dcc9d6d5-dlww8" Feb 02 15:21:34 crc kubenswrapper[4733]: I0202 15:21:34.269205 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-smvdc" Feb 02 15:21:34 crc kubenswrapper[4733]: I0202 15:21:34.269452 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bcccf11d-2567-42d7-81e2-5684b4b43823-utilities\") pod \"redhat-operators-kqbhk\" (UID: \"bcccf11d-2567-42d7-81e2-5684b4b43823\") " pod="openshift-marketplace/redhat-operators-kqbhk" Feb 02 15:21:34 crc kubenswrapper[4733]: I0202 15:21:34.269514 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxbnv\" (UniqueName: \"kubernetes.io/projected/bcccf11d-2567-42d7-81e2-5684b4b43823-kube-api-access-gxbnv\") pod \"redhat-operators-kqbhk\" (UID: \"bcccf11d-2567-42d7-81e2-5684b4b43823\") " pod="openshift-marketplace/redhat-operators-kqbhk" Feb 02 15:21:34 crc kubenswrapper[4733]: I0202 15:21:34.269550 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bcccf11d-2567-42d7-81e2-5684b4b43823-catalog-content\") pod \"redhat-operators-kqbhk\" (UID: \"bcccf11d-2567-42d7-81e2-5684b4b43823\") " pod="openshift-marketplace/redhat-operators-kqbhk" Feb 02 15:21:34 crc kubenswrapper[4733]: I0202 15:21:34.369392 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-75dcc9d6d5-dlww8"] Feb 02 15:21:34 crc kubenswrapper[4733]: I0202 15:21:34.371224 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxbnv\" (UniqueName: \"kubernetes.io/projected/bcccf11d-2567-42d7-81e2-5684b4b43823-kube-api-access-gxbnv\") pod \"redhat-operators-kqbhk\" (UID: \"bcccf11d-2567-42d7-81e2-5684b4b43823\") " pod="openshift-marketplace/redhat-operators-kqbhk" Feb 02 15:21:34 crc kubenswrapper[4733]: I0202 15:21:34.371276 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bcccf11d-2567-42d7-81e2-5684b4b43823-catalog-content\") pod \"redhat-operators-kqbhk\" (UID: \"bcccf11d-2567-42d7-81e2-5684b4b43823\") " pod="openshift-marketplace/redhat-operators-kqbhk" Feb 02 15:21:34 crc kubenswrapper[4733]: I0202 15:21:34.371367 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bcccf11d-2567-42d7-81e2-5684b4b43823-utilities\") pod \"redhat-operators-kqbhk\" (UID: \"bcccf11d-2567-42d7-81e2-5684b4b43823\") " pod="openshift-marketplace/redhat-operators-kqbhk" Feb 02 15:21:34 crc kubenswrapper[4733]: I0202 15:21:34.371740 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bcccf11d-2567-42d7-81e2-5684b4b43823-utilities\") pod \"redhat-operators-kqbhk\" (UID: \"bcccf11d-2567-42d7-81e2-5684b4b43823\") " pod="openshift-marketplace/redhat-operators-kqbhk" Feb 02 15:21:34 crc kubenswrapper[4733]: I0202 15:21:34.372211 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bcccf11d-2567-42d7-81e2-5684b4b43823-catalog-content\") pod \"redhat-operators-kqbhk\" (UID: \"bcccf11d-2567-42d7-81e2-5684b4b43823\") " pod="openshift-marketplace/redhat-operators-kqbhk" Feb 02 15:21:34 crc kubenswrapper[4733]: I0202 15:21:34.389307 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxbnv\" (UniqueName: \"kubernetes.io/projected/bcccf11d-2567-42d7-81e2-5684b4b43823-kube-api-access-gxbnv\") pod \"redhat-operators-kqbhk\" (UID: \"bcccf11d-2567-42d7-81e2-5684b4b43823\") " pod="openshift-marketplace/redhat-operators-kqbhk" Feb 02 15:21:34 crc kubenswrapper[4733]: I0202 15:21:34.465221 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-smvdc"] Feb 02 15:21:34 crc kubenswrapper[4733]: W0202 15:21:34.473883 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0367d1cb_4122_4c7e_b203_d5c4216c6ef5.slice/crio-fe75c1271702130b7520db6ff68bb00987160afc3ec2ccf319bca0542ba8deb9 WatchSource:0}: Error finding container fe75c1271702130b7520db6ff68bb00987160afc3ec2ccf319bca0542ba8deb9: Status 404 returned error can't find the container with id fe75c1271702130b7520db6ff68bb00987160afc3ec2ccf319bca0542ba8deb9 Feb 02 15:21:34 crc kubenswrapper[4733]: I0202 15:21:34.516439 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kqbhk" Feb 02 15:21:34 crc kubenswrapper[4733]: I0202 15:21:34.947883 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kqbhk"] Feb 02 15:21:34 crc kubenswrapper[4733]: W0202 15:21:34.954093 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbcccf11d_2567_42d7_81e2_5684b4b43823.slice/crio-34e6f81ad2bb72f583bd768dc7d9de41a4c4bf169210c981ecfcfed779b3547c WatchSource:0}: Error finding container 34e6f81ad2bb72f583bd768dc7d9de41a4c4bf169210c981ecfcfed779b3547c: Status 404 returned error can't find the container with id 34e6f81ad2bb72f583bd768dc7d9de41a4c4bf169210c981ecfcfed779b3547c Feb 02 15:21:34 crc kubenswrapper[4733]: I0202 15:21:34.987968 4733 patch_prober.go:28] interesting pod/machine-config-daemon-zgsv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 15:21:34 crc kubenswrapper[4733]: I0202 15:21:34.988040 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 15:21:34 crc kubenswrapper[4733]: I0202 15:21:34.997429 4733 generic.go:334] "Generic (PLEG): container finished" podID="54e05a7f-a312-4455-83ec-be97bf446b68" containerID="d0259ef30d7e89b3164987c2cfe9dc5bd29abba98ab18ad6caca4259d42b943f" exitCode=0 Feb 02 15:21:34 crc kubenswrapper[4733]: I0202 15:21:34.997513 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qvwpw" event={"ID":"54e05a7f-a312-4455-83ec-be97bf446b68","Type":"ContainerDied","Data":"d0259ef30d7e89b3164987c2cfe9dc5bd29abba98ab18ad6caca4259d42b943f"} Feb 02 15:21:35 crc kubenswrapper[4733]: I0202 15:21:35.000046 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-75dcc9d6d5-dlww8" event={"ID":"1b138594-1c90-48e0-83a4-e80f3608f0d2","Type":"ContainerStarted","Data":"e964b77eeca011790a4db9127a3b032566c0f42d52d8c20dc63e5fdc7ece4266"} Feb 02 15:21:35 crc kubenswrapper[4733]: I0202 15:21:35.000086 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-75dcc9d6d5-dlww8" event={"ID":"1b138594-1c90-48e0-83a4-e80f3608f0d2","Type":"ContainerStarted","Data":"4c9b2e5d37cc815f4debc5e12f1f29bf2261f743edfd9f5c36cc27615f65a500"} Feb 02 15:21:35 crc kubenswrapper[4733]: I0202 15:21:35.000437 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-75dcc9d6d5-dlww8" Feb 02 15:21:35 crc kubenswrapper[4733]: I0202 15:21:35.006372 4733 generic.go:334] "Generic (PLEG): container finished" podID="b8fd49b0-ac5d-4b93-8c03-7ac35c150aa2" containerID="d42472f1b59d17a8fb8a2f50fb1ad3e5afd2dcfd9a4d7d273df38f547f6fe3e2" exitCode=0 Feb 02 15:21:35 crc kubenswrapper[4733]: I0202 15:21:35.006455 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mbh5r" event={"ID":"b8fd49b0-ac5d-4b93-8c03-7ac35c150aa2","Type":"ContainerDied","Data":"d42472f1b59d17a8fb8a2f50fb1ad3e5afd2dcfd9a4d7d273df38f547f6fe3e2"} Feb 02 15:21:35 crc kubenswrapper[4733]: I0202 15:21:35.017973 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-75dcc9d6d5-dlww8" Feb 02 15:21:35 crc kubenswrapper[4733]: I0202 15:21:35.018113 4733 generic.go:334] "Generic (PLEG): container finished" podID="0367d1cb-4122-4c7e-b203-d5c4216c6ef5" containerID="32259cb39a35f32fcf12b4648afb962ac9a8f64d79f78e74d116a129b7933e29" exitCode=0 Feb 02 15:21:35 crc kubenswrapper[4733]: I0202 15:21:35.018154 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-smvdc" event={"ID":"0367d1cb-4122-4c7e-b203-d5c4216c6ef5","Type":"ContainerDied","Data":"32259cb39a35f32fcf12b4648afb962ac9a8f64d79f78e74d116a129b7933e29"} Feb 02 15:21:35 crc kubenswrapper[4733]: I0202 15:21:35.018190 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-smvdc" event={"ID":"0367d1cb-4122-4c7e-b203-d5c4216c6ef5","Type":"ContainerStarted","Data":"fe75c1271702130b7520db6ff68bb00987160afc3ec2ccf319bca0542ba8deb9"} Feb 02 15:21:35 crc kubenswrapper[4733]: I0202 15:21:35.022454 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kqbhk" event={"ID":"bcccf11d-2567-42d7-81e2-5684b4b43823","Type":"ContainerStarted","Data":"34e6f81ad2bb72f583bd768dc7d9de41a4c4bf169210c981ecfcfed779b3547c"} Feb 02 15:21:35 crc kubenswrapper[4733]: I0202 15:21:35.071068 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-75dcc9d6d5-dlww8" podStartSLOduration=3.071050611 podStartE2EDuration="3.071050611s" podCreationTimestamp="2026-02-02 15:21:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:21:35.067766787 +0000 UTC m=+378.519228145" watchObservedRunningTime="2026-02-02 15:21:35.071050611 +0000 UTC m=+378.522511969" Feb 02 15:21:35 crc kubenswrapper[4733]: I0202 15:21:35.806234 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-brfn5" Feb 02 15:21:35 crc kubenswrapper[4733]: I0202 15:21:35.855534 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-hpgjp"] Feb 02 15:21:36 crc kubenswrapper[4733]: I0202 15:21:36.031352 4733 generic.go:334] "Generic (PLEG): container finished" podID="0367d1cb-4122-4c7e-b203-d5c4216c6ef5" containerID="f86dd0857f58fbbc35619309657cecdd6f5b31f6fcec742fa0b06de038dea124" exitCode=0 Feb 02 15:21:36 crc kubenswrapper[4733]: I0202 15:21:36.031413 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-smvdc" event={"ID":"0367d1cb-4122-4c7e-b203-d5c4216c6ef5","Type":"ContainerDied","Data":"f86dd0857f58fbbc35619309657cecdd6f5b31f6fcec742fa0b06de038dea124"} Feb 02 15:21:36 crc kubenswrapper[4733]: I0202 15:21:36.034348 4733 generic.go:334] "Generic (PLEG): container finished" podID="bcccf11d-2567-42d7-81e2-5684b4b43823" containerID="4d35e63d749072a60edf81789d7e3c9d488e7814d41acda463333e0a44f4fb99" exitCode=0 Feb 02 15:21:36 crc kubenswrapper[4733]: I0202 15:21:36.034598 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kqbhk" event={"ID":"bcccf11d-2567-42d7-81e2-5684b4b43823","Type":"ContainerDied","Data":"4d35e63d749072a60edf81789d7e3c9d488e7814d41acda463333e0a44f4fb99"} Feb 02 15:21:36 crc kubenswrapper[4733]: I0202 15:21:36.040807 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qvwpw" event={"ID":"54e05a7f-a312-4455-83ec-be97bf446b68","Type":"ContainerStarted","Data":"d2cc28fa83c02d2278972d8e115821ced8552a468365dc57ccd647c637e16907"} Feb 02 15:21:36 crc kubenswrapper[4733]: I0202 15:21:36.043065 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mbh5r" event={"ID":"b8fd49b0-ac5d-4b93-8c03-7ac35c150aa2","Type":"ContainerStarted","Data":"f27702b2951d4e92fa479b20b5df9acba1dd4ed585ef79e9b18ed5996d896859"} Feb 02 15:21:36 crc kubenswrapper[4733]: I0202 15:21:36.078112 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mbh5r" podStartSLOduration=2.566547692 podStartE2EDuration="5.078096077s" podCreationTimestamp="2026-02-02 15:21:31 +0000 UTC" firstStartedPulling="2026-02-02 15:21:32.979876915 +0000 UTC m=+376.431338273" lastFinishedPulling="2026-02-02 15:21:35.4914253 +0000 UTC m=+378.942886658" observedRunningTime="2026-02-02 15:21:36.07644271 +0000 UTC m=+379.527904068" watchObservedRunningTime="2026-02-02 15:21:36.078096077 +0000 UTC m=+379.529557435" Feb 02 15:21:36 crc kubenswrapper[4733]: I0202 15:21:36.101602 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-qvwpw" podStartSLOduration=2.661239033 podStartE2EDuration="5.101586734s" podCreationTimestamp="2026-02-02 15:21:31 +0000 UTC" firstStartedPulling="2026-02-02 15:21:32.970748486 +0000 UTC m=+376.422209844" lastFinishedPulling="2026-02-02 15:21:35.411096187 +0000 UTC m=+378.862557545" observedRunningTime="2026-02-02 15:21:36.100334059 +0000 UTC m=+379.551795417" watchObservedRunningTime="2026-02-02 15:21:36.101586734 +0000 UTC m=+379.553048092" Feb 02 15:21:37 crc kubenswrapper[4733]: I0202 15:21:37.050191 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-smvdc" event={"ID":"0367d1cb-4122-4c7e-b203-d5c4216c6ef5","Type":"ContainerStarted","Data":"205b781c92a7ecd28843f314fc5dd6b1006ec9306e4daa73a045ed227fdf5a32"} Feb 02 15:21:37 crc kubenswrapper[4733]: I0202 15:21:37.051755 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kqbhk" event={"ID":"bcccf11d-2567-42d7-81e2-5684b4b43823","Type":"ContainerStarted","Data":"bbd9d76f911a777a4fd354add2c4f2eda8cdc4cd6c10d8beb45facbc2c589f72"} Feb 02 15:21:37 crc kubenswrapper[4733]: I0202 15:21:37.070849 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-smvdc" podStartSLOduration=2.615996762 podStartE2EDuration="4.070829717s" podCreationTimestamp="2026-02-02 15:21:33 +0000 UTC" firstStartedPulling="2026-02-02 15:21:35.019566597 +0000 UTC m=+378.471027955" lastFinishedPulling="2026-02-02 15:21:36.474399552 +0000 UTC m=+379.925860910" observedRunningTime="2026-02-02 15:21:37.065941628 +0000 UTC m=+380.517402996" watchObservedRunningTime="2026-02-02 15:21:37.070829717 +0000 UTC m=+380.522291075" Feb 02 15:21:38 crc kubenswrapper[4733]: I0202 15:21:38.058732 4733 generic.go:334] "Generic (PLEG): container finished" podID="bcccf11d-2567-42d7-81e2-5684b4b43823" containerID="bbd9d76f911a777a4fd354add2c4f2eda8cdc4cd6c10d8beb45facbc2c589f72" exitCode=0 Feb 02 15:21:38 crc kubenswrapper[4733]: I0202 15:21:38.059874 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kqbhk" event={"ID":"bcccf11d-2567-42d7-81e2-5684b4b43823","Type":"ContainerDied","Data":"bbd9d76f911a777a4fd354add2c4f2eda8cdc4cd6c10d8beb45facbc2c589f72"} Feb 02 15:21:40 crc kubenswrapper[4733]: I0202 15:21:40.074232 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kqbhk" event={"ID":"bcccf11d-2567-42d7-81e2-5684b4b43823","Type":"ContainerStarted","Data":"b7e71b90f3ee28d93edfe179039195afab09b5fb092bf89a896f3f8d1e1b8d67"} Feb 02 15:21:40 crc kubenswrapper[4733]: I0202 15:21:40.095649 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kqbhk" podStartSLOduration=2.648557263 podStartE2EDuration="6.095627731s" podCreationTimestamp="2026-02-02 15:21:34 +0000 UTC" firstStartedPulling="2026-02-02 15:21:36.039232173 +0000 UTC m=+379.490693521" lastFinishedPulling="2026-02-02 15:21:39.486302591 +0000 UTC m=+382.937763989" observedRunningTime="2026-02-02 15:21:40.090710611 +0000 UTC m=+383.542171979" watchObservedRunningTime="2026-02-02 15:21:40.095627731 +0000 UTC m=+383.547089109" Feb 02 15:21:41 crc kubenswrapper[4733]: I0202 15:21:41.863389 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mbh5r" Feb 02 15:21:41 crc kubenswrapper[4733]: I0202 15:21:41.864718 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mbh5r" Feb 02 15:21:41 crc kubenswrapper[4733]: I0202 15:21:41.906029 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mbh5r" Feb 02 15:21:42 crc kubenswrapper[4733]: I0202 15:21:42.077277 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-qvwpw" Feb 02 15:21:42 crc kubenswrapper[4733]: I0202 15:21:42.077668 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-qvwpw" Feb 02 15:21:42 crc kubenswrapper[4733]: I0202 15:21:42.114573 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-qvwpw" Feb 02 15:21:42 crc kubenswrapper[4733]: I0202 15:21:42.144433 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mbh5r" Feb 02 15:21:43 crc kubenswrapper[4733]: I0202 15:21:43.124500 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-qvwpw" Feb 02 15:21:44 crc kubenswrapper[4733]: I0202 15:21:44.270550 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-smvdc" Feb 02 15:21:44 crc kubenswrapper[4733]: I0202 15:21:44.270626 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-smvdc" Feb 02 15:21:44 crc kubenswrapper[4733]: I0202 15:21:44.320520 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-smvdc" Feb 02 15:21:44 crc kubenswrapper[4733]: I0202 15:21:44.517627 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kqbhk" Feb 02 15:21:44 crc kubenswrapper[4733]: I0202 15:21:44.517673 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kqbhk" Feb 02 15:21:45 crc kubenswrapper[4733]: I0202 15:21:45.138550 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-smvdc" Feb 02 15:21:45 crc kubenswrapper[4733]: I0202 15:21:45.571471 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-kqbhk" podUID="bcccf11d-2567-42d7-81e2-5684b4b43823" containerName="registry-server" probeResult="failure" output=< Feb 02 15:21:45 crc kubenswrapper[4733]: timeout: failed to connect service ":50051" within 1s Feb 02 15:21:45 crc kubenswrapper[4733]: > Feb 02 15:21:54 crc kubenswrapper[4733]: I0202 15:21:54.580790 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kqbhk" Feb 02 15:21:54 crc kubenswrapper[4733]: I0202 15:21:54.652891 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kqbhk" Feb 02 15:22:00 crc kubenswrapper[4733]: I0202 15:22:00.932782 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" podUID="db551bb0-6793-40de-9b5b-c0ee0dc62711" containerName="registry" containerID="cri-o://7980416916a00f1a8903b9ad92053dd587dfeb5605af745c92f56f0db007f90f" gracePeriod=30 Feb 02 15:22:01 crc kubenswrapper[4733]: I0202 15:22:01.201509 4733 generic.go:334] "Generic (PLEG): container finished" podID="db551bb0-6793-40de-9b5b-c0ee0dc62711" containerID="7980416916a00f1a8903b9ad92053dd587dfeb5605af745c92f56f0db007f90f" exitCode=0 Feb 02 15:22:01 crc kubenswrapper[4733]: I0202 15:22:01.201564 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" event={"ID":"db551bb0-6793-40de-9b5b-c0ee0dc62711","Type":"ContainerDied","Data":"7980416916a00f1a8903b9ad92053dd587dfeb5605af745c92f56f0db007f90f"} Feb 02 15:22:01 crc kubenswrapper[4733]: I0202 15:22:01.416004 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:22:01 crc kubenswrapper[4733]: I0202 15:22:01.566655 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/db551bb0-6793-40de-9b5b-c0ee0dc62711-registry-tls\") pod \"db551bb0-6793-40de-9b5b-c0ee0dc62711\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " Feb 02 15:22:01 crc kubenswrapper[4733]: I0202 15:22:01.566766 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/db551bb0-6793-40de-9b5b-c0ee0dc62711-bound-sa-token\") pod \"db551bb0-6793-40de-9b5b-c0ee0dc62711\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " Feb 02 15:22:01 crc kubenswrapper[4733]: I0202 15:22:01.566900 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/db551bb0-6793-40de-9b5b-c0ee0dc62711-ca-trust-extracted\") pod \"db551bb0-6793-40de-9b5b-c0ee0dc62711\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " Feb 02 15:22:01 crc kubenswrapper[4733]: I0202 15:22:01.566955 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mw5h2\" (UniqueName: \"kubernetes.io/projected/db551bb0-6793-40de-9b5b-c0ee0dc62711-kube-api-access-mw5h2\") pod \"db551bb0-6793-40de-9b5b-c0ee0dc62711\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " Feb 02 15:22:01 crc kubenswrapper[4733]: I0202 15:22:01.567008 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/db551bb0-6793-40de-9b5b-c0ee0dc62711-trusted-ca\") pod \"db551bb0-6793-40de-9b5b-c0ee0dc62711\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " Feb 02 15:22:01 crc kubenswrapper[4733]: I0202 15:22:01.567058 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/db551bb0-6793-40de-9b5b-c0ee0dc62711-installation-pull-secrets\") pod \"db551bb0-6793-40de-9b5b-c0ee0dc62711\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " Feb 02 15:22:01 crc kubenswrapper[4733]: I0202 15:22:01.567345 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"db551bb0-6793-40de-9b5b-c0ee0dc62711\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " Feb 02 15:22:01 crc kubenswrapper[4733]: I0202 15:22:01.567418 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/db551bb0-6793-40de-9b5b-c0ee0dc62711-registry-certificates\") pod \"db551bb0-6793-40de-9b5b-c0ee0dc62711\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " Feb 02 15:22:01 crc kubenswrapper[4733]: I0202 15:22:01.568780 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db551bb0-6793-40de-9b5b-c0ee0dc62711-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "db551bb0-6793-40de-9b5b-c0ee0dc62711" (UID: "db551bb0-6793-40de-9b5b-c0ee0dc62711"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:22:01 crc kubenswrapper[4733]: I0202 15:22:01.568844 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db551bb0-6793-40de-9b5b-c0ee0dc62711-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "db551bb0-6793-40de-9b5b-c0ee0dc62711" (UID: "db551bb0-6793-40de-9b5b-c0ee0dc62711"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:22:01 crc kubenswrapper[4733]: I0202 15:22:01.573445 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db551bb0-6793-40de-9b5b-c0ee0dc62711-kube-api-access-mw5h2" (OuterVolumeSpecName: "kube-api-access-mw5h2") pod "db551bb0-6793-40de-9b5b-c0ee0dc62711" (UID: "db551bb0-6793-40de-9b5b-c0ee0dc62711"). InnerVolumeSpecName "kube-api-access-mw5h2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:22:01 crc kubenswrapper[4733]: I0202 15:22:01.575267 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db551bb0-6793-40de-9b5b-c0ee0dc62711-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "db551bb0-6793-40de-9b5b-c0ee0dc62711" (UID: "db551bb0-6793-40de-9b5b-c0ee0dc62711"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:22:01 crc kubenswrapper[4733]: I0202 15:22:01.575499 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db551bb0-6793-40de-9b5b-c0ee0dc62711-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "db551bb0-6793-40de-9b5b-c0ee0dc62711" (UID: "db551bb0-6793-40de-9b5b-c0ee0dc62711"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:22:01 crc kubenswrapper[4733]: E0202 15:22:01.578470 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:db551bb0-6793-40de-9b5b-c0ee0dc62711 nodeName:}" failed. No retries permitted until 2026-02-02 15:22:02.078430055 +0000 UTC m=+405.529891453 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "registry-storage" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "db551bb0-6793-40de-9b5b-c0ee0dc62711" (UID: "db551bb0-6793-40de-9b5b-c0ee0dc62711") : kubernetes.io/csi: Unmounter.TearDownAt failed: rpc error: code = Unknown desc = check target path: could not get consistent content of /proc/mounts after 3 attempts Feb 02 15:22:01 crc kubenswrapper[4733]: I0202 15:22:01.579649 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db551bb0-6793-40de-9b5b-c0ee0dc62711-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "db551bb0-6793-40de-9b5b-c0ee0dc62711" (UID: "db551bb0-6793-40de-9b5b-c0ee0dc62711"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:22:01 crc kubenswrapper[4733]: I0202 15:22:01.588210 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db551bb0-6793-40de-9b5b-c0ee0dc62711-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "db551bb0-6793-40de-9b5b-c0ee0dc62711" (UID: "db551bb0-6793-40de-9b5b-c0ee0dc62711"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:22:01 crc kubenswrapper[4733]: I0202 15:22:01.669565 4733 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/db551bb0-6793-40de-9b5b-c0ee0dc62711-registry-tls\") on node \"crc\" DevicePath \"\"" Feb 02 15:22:01 crc kubenswrapper[4733]: I0202 15:22:01.669618 4733 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/db551bb0-6793-40de-9b5b-c0ee0dc62711-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 02 15:22:01 crc kubenswrapper[4733]: I0202 15:22:01.669639 4733 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/db551bb0-6793-40de-9b5b-c0ee0dc62711-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Feb 02 15:22:01 crc kubenswrapper[4733]: I0202 15:22:01.669661 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mw5h2\" (UniqueName: \"kubernetes.io/projected/db551bb0-6793-40de-9b5b-c0ee0dc62711-kube-api-access-mw5h2\") on node \"crc\" DevicePath \"\"" Feb 02 15:22:01 crc kubenswrapper[4733]: I0202 15:22:01.669680 4733 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/db551bb0-6793-40de-9b5b-c0ee0dc62711-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 02 15:22:01 crc kubenswrapper[4733]: I0202 15:22:01.669699 4733 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/db551bb0-6793-40de-9b5b-c0ee0dc62711-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Feb 02 15:22:01 crc kubenswrapper[4733]: I0202 15:22:01.669719 4733 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/db551bb0-6793-40de-9b5b-c0ee0dc62711-registry-certificates\") on node \"crc\" DevicePath \"\"" Feb 02 15:22:02 crc kubenswrapper[4733]: I0202 15:22:02.176420 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"db551bb0-6793-40de-9b5b-c0ee0dc62711\" (UID: \"db551bb0-6793-40de-9b5b-c0ee0dc62711\") " Feb 02 15:22:02 crc kubenswrapper[4733]: I0202 15:22:02.188746 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "db551bb0-6793-40de-9b5b-c0ee0dc62711" (UID: "db551bb0-6793-40de-9b5b-c0ee0dc62711"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 02 15:22:02 crc kubenswrapper[4733]: I0202 15:22:02.209649 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" event={"ID":"db551bb0-6793-40de-9b5b-c0ee0dc62711","Type":"ContainerDied","Data":"29f18ceaa31f8bf05fff8d43cd439a5b7380e1ea586c593065c2aa7a1adf1de2"} Feb 02 15:22:02 crc kubenswrapper[4733]: I0202 15:22:02.209703 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-hpgjp" Feb 02 15:22:02 crc kubenswrapper[4733]: I0202 15:22:02.209738 4733 scope.go:117] "RemoveContainer" containerID="7980416916a00f1a8903b9ad92053dd587dfeb5605af745c92f56f0db007f90f" Feb 02 15:22:02 crc kubenswrapper[4733]: I0202 15:22:02.249634 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-hpgjp"] Feb 02 15:22:02 crc kubenswrapper[4733]: I0202 15:22:02.253255 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-hpgjp"] Feb 02 15:22:02 crc kubenswrapper[4733]: E0202 15:22:02.264764 4733 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddb551bb0_6793_40de_9b5b_c0ee0dc62711.slice\": RecentStats: unable to find data in memory cache]" Feb 02 15:22:03 crc kubenswrapper[4733]: I0202 15:22:03.264715 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db551bb0-6793-40de-9b5b-c0ee0dc62711" path="/var/lib/kubelet/pods/db551bb0-6793-40de-9b5b-c0ee0dc62711/volumes" Feb 02 15:22:04 crc kubenswrapper[4733]: I0202 15:22:04.988242 4733 patch_prober.go:28] interesting pod/machine-config-daemon-zgsv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 15:22:04 crc kubenswrapper[4733]: I0202 15:22:04.988326 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 15:22:34 crc kubenswrapper[4733]: I0202 15:22:34.988436 4733 patch_prober.go:28] interesting pod/machine-config-daemon-zgsv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 15:22:34 crc kubenswrapper[4733]: I0202 15:22:34.989363 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 15:22:34 crc kubenswrapper[4733]: I0202 15:22:34.989447 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" Feb 02 15:22:34 crc kubenswrapper[4733]: I0202 15:22:34.990282 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c5ef374155c9a13227e85afc96d240c8534badb0e00280dca27672b354c63947"} pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 15:22:34 crc kubenswrapper[4733]: I0202 15:22:34.990371 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" containerName="machine-config-daemon" containerID="cri-o://c5ef374155c9a13227e85afc96d240c8534badb0e00280dca27672b354c63947" gracePeriod=600 Feb 02 15:22:35 crc kubenswrapper[4733]: I0202 15:22:35.467349 4733 generic.go:334] "Generic (PLEG): container finished" podID="83d0e810-ea5c-47aa-947b-7865b045bf94" containerID="c5ef374155c9a13227e85afc96d240c8534badb0e00280dca27672b354c63947" exitCode=0 Feb 02 15:22:35 crc kubenswrapper[4733]: I0202 15:22:35.467435 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" event={"ID":"83d0e810-ea5c-47aa-947b-7865b045bf94","Type":"ContainerDied","Data":"c5ef374155c9a13227e85afc96d240c8534badb0e00280dca27672b354c63947"} Feb 02 15:22:35 crc kubenswrapper[4733]: I0202 15:22:35.467724 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" event={"ID":"83d0e810-ea5c-47aa-947b-7865b045bf94","Type":"ContainerStarted","Data":"6cb02659ae49ba1eb2bfa677c715cc0aaf1abb6c83347c16358782946f3f3f0f"} Feb 02 15:22:35 crc kubenswrapper[4733]: I0202 15:22:35.467770 4733 scope.go:117] "RemoveContainer" containerID="d51318ad71b42f4cc8d376aa50e44580e68da89d83d80defb996329a2ee5c2ad" Feb 02 15:25:04 crc kubenswrapper[4733]: I0202 15:25:04.987556 4733 patch_prober.go:28] interesting pod/machine-config-daemon-zgsv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 15:25:04 crc kubenswrapper[4733]: I0202 15:25:04.988243 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 15:25:34 crc kubenswrapper[4733]: I0202 15:25:34.987732 4733 patch_prober.go:28] interesting pod/machine-config-daemon-zgsv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 15:25:34 crc kubenswrapper[4733]: I0202 15:25:34.988458 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 15:26:04 crc kubenswrapper[4733]: I0202 15:26:04.988114 4733 patch_prober.go:28] interesting pod/machine-config-daemon-zgsv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 15:26:04 crc kubenswrapper[4733]: I0202 15:26:04.988865 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 15:26:04 crc kubenswrapper[4733]: I0202 15:26:04.988933 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" Feb 02 15:26:04 crc kubenswrapper[4733]: I0202 15:26:04.989796 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6cb02659ae49ba1eb2bfa677c715cc0aaf1abb6c83347c16358782946f3f3f0f"} pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 15:26:04 crc kubenswrapper[4733]: I0202 15:26:04.989877 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" containerName="machine-config-daemon" containerID="cri-o://6cb02659ae49ba1eb2bfa677c715cc0aaf1abb6c83347c16358782946f3f3f0f" gracePeriod=600 Feb 02 15:26:05 crc kubenswrapper[4733]: I0202 15:26:05.855270 4733 generic.go:334] "Generic (PLEG): container finished" podID="83d0e810-ea5c-47aa-947b-7865b045bf94" containerID="6cb02659ae49ba1eb2bfa677c715cc0aaf1abb6c83347c16358782946f3f3f0f" exitCode=0 Feb 02 15:26:05 crc kubenswrapper[4733]: I0202 15:26:05.855314 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" event={"ID":"83d0e810-ea5c-47aa-947b-7865b045bf94","Type":"ContainerDied","Data":"6cb02659ae49ba1eb2bfa677c715cc0aaf1abb6c83347c16358782946f3f3f0f"} Feb 02 15:26:05 crc kubenswrapper[4733]: I0202 15:26:05.855921 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" event={"ID":"83d0e810-ea5c-47aa-947b-7865b045bf94","Type":"ContainerStarted","Data":"25cacf7f1a9153fe0cadb8062fa48df6b9242dcf030528fe627cf68ad889d6b1"} Feb 02 15:26:05 crc kubenswrapper[4733]: I0202 15:26:05.855955 4733 scope.go:117] "RemoveContainer" containerID="c5ef374155c9a13227e85afc96d240c8534badb0e00280dca27672b354c63947" Feb 02 15:26:40 crc kubenswrapper[4733]: I0202 15:26:40.327822 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qxvkf"] Feb 02 15:26:40 crc kubenswrapper[4733]: I0202 15:26:40.329134 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" containerName="ovn-controller" containerID="cri-o://db1db75e392ebea8b387b85cabc4276d553bf2c06694cf7f294a383909f32826" gracePeriod=30 Feb 02 15:26:40 crc kubenswrapper[4733]: I0202 15:26:40.329351 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://831d85055203f153381eb51d98af0cd2e4c4a0309f837545ab3d855a680133e2" gracePeriod=30 Feb 02 15:26:40 crc kubenswrapper[4733]: I0202 15:26:40.329312 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" containerName="northd" containerID="cri-o://c5835cb53385359e9c58eaa5d4da2569c313e0fc265ad79a6c124e8cecb1a734" gracePeriod=30 Feb 02 15:26:40 crc kubenswrapper[4733]: I0202 15:26:40.329424 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" containerName="ovn-acl-logging" containerID="cri-o://ffd6dc4bf5b9fd31bfbee6d9340fe981c0c5fca03a79cef9316340ff5a8b3d0c" gracePeriod=30 Feb 02 15:26:40 crc kubenswrapper[4733]: I0202 15:26:40.329334 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" containerName="nbdb" containerID="cri-o://1380122d1694c1286f1102e6396c6504b91f63c6eb157689dfc0ea7188c17f68" gracePeriod=30 Feb 02 15:26:40 crc kubenswrapper[4733]: I0202 15:26:40.329599 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" containerName="kube-rbac-proxy-node" containerID="cri-o://4acf2398e2e19bd54f482cb2273c54738e61627187860bb0b6daadddc2565b63" gracePeriod=30 Feb 02 15:26:40 crc kubenswrapper[4733]: I0202 15:26:40.329704 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" containerName="sbdb" containerID="cri-o://6efaf2216e4c1750139f080a93b4b608737dab2883de5fddfd9a99248d517b2c" gracePeriod=30 Feb 02 15:26:40 crc kubenswrapper[4733]: I0202 15:26:40.401923 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" containerName="ovnkube-controller" containerID="cri-o://21363125c857d52144029dc46a3a1553d0b588546a6aaeab91155f3802857fc5" gracePeriod=30 Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.081523 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxvkf_f5a81ada-e393-499a-aac6-7b40349958bb/ovnkube-controller/3.log" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.086199 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxvkf_f5a81ada-e393-499a-aac6-7b40349958bb/ovn-acl-logging/0.log" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.087203 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxvkf_f5a81ada-e393-499a-aac6-7b40349958bb/ovn-controller/0.log" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.087904 4733 generic.go:334] "Generic (PLEG): container finished" podID="f5a81ada-e393-499a-aac6-7b40349958bb" containerID="21363125c857d52144029dc46a3a1553d0b588546a6aaeab91155f3802857fc5" exitCode=0 Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.087949 4733 generic.go:334] "Generic (PLEG): container finished" podID="f5a81ada-e393-499a-aac6-7b40349958bb" containerID="6efaf2216e4c1750139f080a93b4b608737dab2883de5fddfd9a99248d517b2c" exitCode=0 Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.087975 4733 generic.go:334] "Generic (PLEG): container finished" podID="f5a81ada-e393-499a-aac6-7b40349958bb" containerID="1380122d1694c1286f1102e6396c6504b91f63c6eb157689dfc0ea7188c17f68" exitCode=0 Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.087992 4733 generic.go:334] "Generic (PLEG): container finished" podID="f5a81ada-e393-499a-aac6-7b40349958bb" containerID="c5835cb53385359e9c58eaa5d4da2569c313e0fc265ad79a6c124e8cecb1a734" exitCode=0 Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.088010 4733 generic.go:334] "Generic (PLEG): container finished" podID="f5a81ada-e393-499a-aac6-7b40349958bb" containerID="831d85055203f153381eb51d98af0cd2e4c4a0309f837545ab3d855a680133e2" exitCode=0 Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.088011 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" event={"ID":"f5a81ada-e393-499a-aac6-7b40349958bb","Type":"ContainerDied","Data":"21363125c857d52144029dc46a3a1553d0b588546a6aaeab91155f3802857fc5"} Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.088087 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" event={"ID":"f5a81ada-e393-499a-aac6-7b40349958bb","Type":"ContainerDied","Data":"6efaf2216e4c1750139f080a93b4b608737dab2883de5fddfd9a99248d517b2c"} Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.088112 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" event={"ID":"f5a81ada-e393-499a-aac6-7b40349958bb","Type":"ContainerDied","Data":"1380122d1694c1286f1102e6396c6504b91f63c6eb157689dfc0ea7188c17f68"} Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.088131 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" event={"ID":"f5a81ada-e393-499a-aac6-7b40349958bb","Type":"ContainerDied","Data":"c5835cb53385359e9c58eaa5d4da2569c313e0fc265ad79a6c124e8cecb1a734"} Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.088138 4733 scope.go:117] "RemoveContainer" containerID="55b2c8f5f7c8b659a1b54bb592229f81596804f972fc4658e1f867054721257d" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.088152 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" event={"ID":"f5a81ada-e393-499a-aac6-7b40349958bb","Type":"ContainerDied","Data":"831d85055203f153381eb51d98af0cd2e4c4a0309f837545ab3d855a680133e2"} Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.088329 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" event={"ID":"f5a81ada-e393-499a-aac6-7b40349958bb","Type":"ContainerDied","Data":"4acf2398e2e19bd54f482cb2273c54738e61627187860bb0b6daadddc2565b63"} Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.088027 4733 generic.go:334] "Generic (PLEG): container finished" podID="f5a81ada-e393-499a-aac6-7b40349958bb" containerID="4acf2398e2e19bd54f482cb2273c54738e61627187860bb0b6daadddc2565b63" exitCode=0 Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.088387 4733 generic.go:334] "Generic (PLEG): container finished" podID="f5a81ada-e393-499a-aac6-7b40349958bb" containerID="ffd6dc4bf5b9fd31bfbee6d9340fe981c0c5fca03a79cef9316340ff5a8b3d0c" exitCode=143 Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.088410 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" event={"ID":"f5a81ada-e393-499a-aac6-7b40349958bb","Type":"ContainerDied","Data":"ffd6dc4bf5b9fd31bfbee6d9340fe981c0c5fca03a79cef9316340ff5a8b3d0c"} Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.088438 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" event={"ID":"f5a81ada-e393-499a-aac6-7b40349958bb","Type":"ContainerDied","Data":"db1db75e392ebea8b387b85cabc4276d553bf2c06694cf7f294a383909f32826"} Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.088417 4733 generic.go:334] "Generic (PLEG): container finished" podID="f5a81ada-e393-499a-aac6-7b40349958bb" containerID="db1db75e392ebea8b387b85cabc4276d553bf2c06694cf7f294a383909f32826" exitCode=143 Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.092471 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jh6vk_b76a05d6-d9c7-4183-ad90-b3131f225841/kube-multus/2.log" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.093345 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jh6vk_b76a05d6-d9c7-4183-ad90-b3131f225841/kube-multus/1.log" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.093429 4733 generic.go:334] "Generic (PLEG): container finished" podID="b76a05d6-d9c7-4183-ad90-b3131f225841" containerID="0746ffd4e82c3554d4f63c139295a9cedb77073db5dda7ba15e32d519c0ed868" exitCode=2 Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.093475 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jh6vk" event={"ID":"b76a05d6-d9c7-4183-ad90-b3131f225841","Type":"ContainerDied","Data":"0746ffd4e82c3554d4f63c139295a9cedb77073db5dda7ba15e32d519c0ed868"} Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.094360 4733 scope.go:117] "RemoveContainer" containerID="0746ffd4e82c3554d4f63c139295a9cedb77073db5dda7ba15e32d519c0ed868" Feb 02 15:26:41 crc kubenswrapper[4733]: E0202 15:26:41.094855 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-jh6vk_openshift-multus(b76a05d6-d9c7-4183-ad90-b3131f225841)\"" pod="openshift-multus/multus-jh6vk" podUID="b76a05d6-d9c7-4183-ad90-b3131f225841" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.191254 4733 scope.go:117] "RemoveContainer" containerID="f57aea989227fdbd7bb59c51fade50d1076929f6288d060808b51bd1bdd7df04" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.217500 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxvkf_f5a81ada-e393-499a-aac6-7b40349958bb/ovn-acl-logging/0.log" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.218315 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxvkf_f5a81ada-e393-499a-aac6-7b40349958bb/ovn-controller/0.log" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.218878 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.252320 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f5a81ada-e393-499a-aac6-7b40349958bb-ovnkube-config\") pod \"f5a81ada-e393-499a-aac6-7b40349958bb\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.252384 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-host-cni-netd\") pod \"f5a81ada-e393-499a-aac6-7b40349958bb\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.252435 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r4z57\" (UniqueName: \"kubernetes.io/projected/f5a81ada-e393-499a-aac6-7b40349958bb-kube-api-access-r4z57\") pod \"f5a81ada-e393-499a-aac6-7b40349958bb\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.252469 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-host-kubelet\") pod \"f5a81ada-e393-499a-aac6-7b40349958bb\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.252499 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-host-run-netns\") pod \"f5a81ada-e393-499a-aac6-7b40349958bb\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.252533 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-etc-openvswitch\") pod \"f5a81ada-e393-499a-aac6-7b40349958bb\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.252564 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f5a81ada-e393-499a-aac6-7b40349958bb-ovn-node-metrics-cert\") pod \"f5a81ada-e393-499a-aac6-7b40349958bb\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.252595 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-host-slash\") pod \"f5a81ada-e393-499a-aac6-7b40349958bb\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.252639 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-systemd-units\") pod \"f5a81ada-e393-499a-aac6-7b40349958bb\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.252669 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-run-ovn\") pod \"f5a81ada-e393-499a-aac6-7b40349958bb\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.252702 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-run-systemd\") pod \"f5a81ada-e393-499a-aac6-7b40349958bb\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.252752 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-var-lib-openvswitch\") pod \"f5a81ada-e393-499a-aac6-7b40349958bb\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.252789 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-run-openvswitch\") pod \"f5a81ada-e393-499a-aac6-7b40349958bb\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.252817 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-host-run-ovn-kubernetes\") pod \"f5a81ada-e393-499a-aac6-7b40349958bb\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.252817 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "f5a81ada-e393-499a-aac6-7b40349958bb" (UID: "f5a81ada-e393-499a-aac6-7b40349958bb"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.252862 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-host-var-lib-cni-networks-ovn-kubernetes\") pod \"f5a81ada-e393-499a-aac6-7b40349958bb\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.252907 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f5a81ada-e393-499a-aac6-7b40349958bb-ovnkube-script-lib\") pod \"f5a81ada-e393-499a-aac6-7b40349958bb\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.252909 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "f5a81ada-e393-499a-aac6-7b40349958bb" (UID: "f5a81ada-e393-499a-aac6-7b40349958bb"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.252970 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-log-socket\") pod \"f5a81ada-e393-499a-aac6-7b40349958bb\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.253003 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f5a81ada-e393-499a-aac6-7b40349958bb-env-overrides\") pod \"f5a81ada-e393-499a-aac6-7b40349958bb\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.253050 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-host-cni-bin\") pod \"f5a81ada-e393-499a-aac6-7b40349958bb\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.253081 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-node-log\") pod \"f5a81ada-e393-499a-aac6-7b40349958bb\" (UID: \"f5a81ada-e393-499a-aac6-7b40349958bb\") " Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.253410 4733 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-host-cni-netd\") on node \"crc\" DevicePath \"\"" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.253433 4733 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.253072 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "f5a81ada-e393-499a-aac6-7b40349958bb" (UID: "f5a81ada-e393-499a-aac6-7b40349958bb"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.253107 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "f5a81ada-e393-499a-aac6-7b40349958bb" (UID: "f5a81ada-e393-499a-aac6-7b40349958bb"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.253465 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5a81ada-e393-499a-aac6-7b40349958bb-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "f5a81ada-e393-499a-aac6-7b40349958bb" (UID: "f5a81ada-e393-499a-aac6-7b40349958bb"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.253127 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "f5a81ada-e393-499a-aac6-7b40349958bb" (UID: "f5a81ada-e393-499a-aac6-7b40349958bb"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.253785 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "f5a81ada-e393-499a-aac6-7b40349958bb" (UID: "f5a81ada-e393-499a-aac6-7b40349958bb"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.253800 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "f5a81ada-e393-499a-aac6-7b40349958bb" (UID: "f5a81ada-e393-499a-aac6-7b40349958bb"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.253153 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-log-socket" (OuterVolumeSpecName: "log-socket") pod "f5a81ada-e393-499a-aac6-7b40349958bb" (UID: "f5a81ada-e393-499a-aac6-7b40349958bb"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.253189 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "f5a81ada-e393-499a-aac6-7b40349958bb" (UID: "f5a81ada-e393-499a-aac6-7b40349958bb"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.253748 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5a81ada-e393-499a-aac6-7b40349958bb-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "f5a81ada-e393-499a-aac6-7b40349958bb" (UID: "f5a81ada-e393-499a-aac6-7b40349958bb"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.253904 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "f5a81ada-e393-499a-aac6-7b40349958bb" (UID: "f5a81ada-e393-499a-aac6-7b40349958bb"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.253949 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-node-log" (OuterVolumeSpecName: "node-log") pod "f5a81ada-e393-499a-aac6-7b40349958bb" (UID: "f5a81ada-e393-499a-aac6-7b40349958bb"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.253845 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-host-slash" (OuterVolumeSpecName: "host-slash") pod "f5a81ada-e393-499a-aac6-7b40349958bb" (UID: "f5a81ada-e393-499a-aac6-7b40349958bb"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.254127 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5a81ada-e393-499a-aac6-7b40349958bb-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "f5a81ada-e393-499a-aac6-7b40349958bb" (UID: "f5a81ada-e393-499a-aac6-7b40349958bb"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.254349 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "f5a81ada-e393-499a-aac6-7b40349958bb" (UID: "f5a81ada-e393-499a-aac6-7b40349958bb"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.254493 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "f5a81ada-e393-499a-aac6-7b40349958bb" (UID: "f5a81ada-e393-499a-aac6-7b40349958bb"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.262564 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5a81ada-e393-499a-aac6-7b40349958bb-kube-api-access-r4z57" (OuterVolumeSpecName: "kube-api-access-r4z57") pod "f5a81ada-e393-499a-aac6-7b40349958bb" (UID: "f5a81ada-e393-499a-aac6-7b40349958bb"). InnerVolumeSpecName "kube-api-access-r4z57". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.270704 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5a81ada-e393-499a-aac6-7b40349958bb-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "f5a81ada-e393-499a-aac6-7b40349958bb" (UID: "f5a81ada-e393-499a-aac6-7b40349958bb"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.285863 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "f5a81ada-e393-499a-aac6-7b40349958bb" (UID: "f5a81ada-e393-499a-aac6-7b40349958bb"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.297803 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-l5ljn"] Feb 02 15:26:41 crc kubenswrapper[4733]: E0202 15:26:41.298083 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" containerName="ovnkube-controller" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.298105 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" containerName="ovnkube-controller" Feb 02 15:26:41 crc kubenswrapper[4733]: E0202 15:26:41.298119 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" containerName="ovnkube-controller" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.298128 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" containerName="ovnkube-controller" Feb 02 15:26:41 crc kubenswrapper[4733]: E0202 15:26:41.298144 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" containerName="kube-rbac-proxy-ovn-metrics" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.298153 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" containerName="kube-rbac-proxy-ovn-metrics" Feb 02 15:26:41 crc kubenswrapper[4733]: E0202 15:26:41.298190 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" containerName="nbdb" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.298199 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" containerName="nbdb" Feb 02 15:26:41 crc kubenswrapper[4733]: E0202 15:26:41.298212 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" containerName="ovn-controller" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.298221 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" containerName="ovn-controller" Feb 02 15:26:41 crc kubenswrapper[4733]: E0202 15:26:41.298232 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" containerName="kube-rbac-proxy-node" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.298242 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" containerName="kube-rbac-proxy-node" Feb 02 15:26:41 crc kubenswrapper[4733]: E0202 15:26:41.298259 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db551bb0-6793-40de-9b5b-c0ee0dc62711" containerName="registry" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.298268 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="db551bb0-6793-40de-9b5b-c0ee0dc62711" containerName="registry" Feb 02 15:26:41 crc kubenswrapper[4733]: E0202 15:26:41.298279 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" containerName="kubecfg-setup" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.298288 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" containerName="kubecfg-setup" Feb 02 15:26:41 crc kubenswrapper[4733]: E0202 15:26:41.298301 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" containerName="ovn-acl-logging" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.298309 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" containerName="ovn-acl-logging" Feb 02 15:26:41 crc kubenswrapper[4733]: E0202 15:26:41.298322 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" containerName="northd" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.298330 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" containerName="northd" Feb 02 15:26:41 crc kubenswrapper[4733]: E0202 15:26:41.298346 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" containerName="sbdb" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.298356 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" containerName="sbdb" Feb 02 15:26:41 crc kubenswrapper[4733]: E0202 15:26:41.298373 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" containerName="ovnkube-controller" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.298384 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" containerName="ovnkube-controller" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.298528 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" containerName="ovn-controller" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.298550 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" containerName="ovnkube-controller" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.298562 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" containerName="sbdb" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.298573 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" containerName="ovnkube-controller" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.298586 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" containerName="nbdb" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.298596 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" containerName="kube-rbac-proxy-ovn-metrics" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.298607 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" containerName="northd" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.298620 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" containerName="kube-rbac-proxy-node" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.298633 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" containerName="ovn-acl-logging" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.298643 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" containerName="ovnkube-controller" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.298657 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="db551bb0-6793-40de-9b5b-c0ee0dc62711" containerName="registry" Feb 02 15:26:41 crc kubenswrapper[4733]: E0202 15:26:41.298770 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" containerName="ovnkube-controller" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.298780 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" containerName="ovnkube-controller" Feb 02 15:26:41 crc kubenswrapper[4733]: E0202 15:26:41.298791 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" containerName="ovnkube-controller" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.298800 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" containerName="ovnkube-controller" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.298924 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" containerName="ovnkube-controller" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.298935 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" containerName="ovnkube-controller" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.301098 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.353923 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/7205a259-4e48-4108-bb99-5d1eb36ca4a1-systemd-units\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.354032 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7205a259-4e48-4108-bb99-5d1eb36ca4a1-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.354121 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7205a259-4e48-4108-bb99-5d1eb36ca4a1-host-run-netns\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.354219 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7205a259-4e48-4108-bb99-5d1eb36ca4a1-run-openvswitch\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.354300 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7205a259-4e48-4108-bb99-5d1eb36ca4a1-etc-openvswitch\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.354333 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8nxs\" (UniqueName: \"kubernetes.io/projected/7205a259-4e48-4108-bb99-5d1eb36ca4a1-kube-api-access-r8nxs\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.354387 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7205a259-4e48-4108-bb99-5d1eb36ca4a1-host-cni-bin\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.354446 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/7205a259-4e48-4108-bb99-5d1eb36ca4a1-run-systemd\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.354518 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7205a259-4e48-4108-bb99-5d1eb36ca4a1-ovn-node-metrics-cert\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.354546 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7205a259-4e48-4108-bb99-5d1eb36ca4a1-ovnkube-config\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.354781 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/7205a259-4e48-4108-bb99-5d1eb36ca4a1-host-slash\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.354893 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/7205a259-4e48-4108-bb99-5d1eb36ca4a1-run-ovn\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.354950 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7205a259-4e48-4108-bb99-5d1eb36ca4a1-var-lib-openvswitch\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.354974 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/7205a259-4e48-4108-bb99-5d1eb36ca4a1-log-socket\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.354999 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7205a259-4e48-4108-bb99-5d1eb36ca4a1-env-overrides\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.355031 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7205a259-4e48-4108-bb99-5d1eb36ca4a1-host-run-ovn-kubernetes\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.355120 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/7205a259-4e48-4108-bb99-5d1eb36ca4a1-host-kubelet\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.355194 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/7205a259-4e48-4108-bb99-5d1eb36ca4a1-host-cni-netd\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.355241 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/7205a259-4e48-4108-bb99-5d1eb36ca4a1-ovnkube-script-lib\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.355271 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/7205a259-4e48-4108-bb99-5d1eb36ca4a1-node-log\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.355342 4733 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f5a81ada-e393-499a-aac6-7b40349958bb-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.355362 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r4z57\" (UniqueName: \"kubernetes.io/projected/f5a81ada-e393-499a-aac6-7b40349958bb-kube-api-access-r4z57\") on node \"crc\" DevicePath \"\"" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.355376 4733 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-host-kubelet\") on node \"crc\" DevicePath \"\"" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.355389 4733 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-host-run-netns\") on node \"crc\" DevicePath \"\"" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.355401 4733 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f5a81ada-e393-499a-aac6-7b40349958bb-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.355413 4733 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-host-slash\") on node \"crc\" DevicePath \"\"" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.355428 4733 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-systemd-units\") on node \"crc\" DevicePath \"\"" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.355441 4733 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-run-ovn\") on node \"crc\" DevicePath \"\"" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.355454 4733 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-run-systemd\") on node \"crc\" DevicePath \"\"" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.355465 4733 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.355477 4733 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-run-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.355489 4733 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.355503 4733 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.355526 4733 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f5a81ada-e393-499a-aac6-7b40349958bb-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.355543 4733 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-log-socket\") on node \"crc\" DevicePath \"\"" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.355557 4733 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f5a81ada-e393-499a-aac6-7b40349958bb-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.355569 4733 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-host-cni-bin\") on node \"crc\" DevicePath \"\"" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.355582 4733 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f5a81ada-e393-499a-aac6-7b40349958bb-node-log\") on node \"crc\" DevicePath \"\"" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.457099 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7205a259-4e48-4108-bb99-5d1eb36ca4a1-ovnkube-config\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.457225 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/7205a259-4e48-4108-bb99-5d1eb36ca4a1-host-slash\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.457265 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/7205a259-4e48-4108-bb99-5d1eb36ca4a1-run-ovn\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.457298 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/7205a259-4e48-4108-bb99-5d1eb36ca4a1-log-socket\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.457319 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7205a259-4e48-4108-bb99-5d1eb36ca4a1-env-overrides\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.457338 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/7205a259-4e48-4108-bb99-5d1eb36ca4a1-host-slash\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.457347 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7205a259-4e48-4108-bb99-5d1eb36ca4a1-var-lib-openvswitch\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.457407 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7205a259-4e48-4108-bb99-5d1eb36ca4a1-var-lib-openvswitch\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.457416 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/7205a259-4e48-4108-bb99-5d1eb36ca4a1-run-ovn\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.457450 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7205a259-4e48-4108-bb99-5d1eb36ca4a1-host-run-ovn-kubernetes\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.457493 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/7205a259-4e48-4108-bb99-5d1eb36ca4a1-host-cni-netd\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.457496 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/7205a259-4e48-4108-bb99-5d1eb36ca4a1-log-socket\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.457543 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/7205a259-4e48-4108-bb99-5d1eb36ca4a1-host-kubelet\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.457519 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/7205a259-4e48-4108-bb99-5d1eb36ca4a1-host-kubelet\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.457525 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7205a259-4e48-4108-bb99-5d1eb36ca4a1-host-run-ovn-kubernetes\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.457569 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/7205a259-4e48-4108-bb99-5d1eb36ca4a1-host-cni-netd\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.457692 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/7205a259-4e48-4108-bb99-5d1eb36ca4a1-ovnkube-script-lib\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.457750 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/7205a259-4e48-4108-bb99-5d1eb36ca4a1-node-log\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.457822 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/7205a259-4e48-4108-bb99-5d1eb36ca4a1-systemd-units\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.457871 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7205a259-4e48-4108-bb99-5d1eb36ca4a1-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.457978 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7205a259-4e48-4108-bb99-5d1eb36ca4a1-host-run-netns\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.458072 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7205a259-4e48-4108-bb99-5d1eb36ca4a1-run-openvswitch\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.458135 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8nxs\" (UniqueName: \"kubernetes.io/projected/7205a259-4e48-4108-bb99-5d1eb36ca4a1-kube-api-access-r8nxs\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.458216 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7205a259-4e48-4108-bb99-5d1eb36ca4a1-etc-openvswitch\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.458280 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7205a259-4e48-4108-bb99-5d1eb36ca4a1-host-cni-bin\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.458338 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/7205a259-4e48-4108-bb99-5d1eb36ca4a1-run-systemd\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.458397 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7205a259-4e48-4108-bb99-5d1eb36ca4a1-ovn-node-metrics-cert\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.458625 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7205a259-4e48-4108-bb99-5d1eb36ca4a1-env-overrides\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.458732 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7205a259-4e48-4108-bb99-5d1eb36ca4a1-run-openvswitch\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.458791 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/7205a259-4e48-4108-bb99-5d1eb36ca4a1-node-log\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.458841 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/7205a259-4e48-4108-bb99-5d1eb36ca4a1-systemd-units\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.458887 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7205a259-4e48-4108-bb99-5d1eb36ca4a1-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.458905 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/7205a259-4e48-4108-bb99-5d1eb36ca4a1-ovnkube-script-lib\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.458921 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7205a259-4e48-4108-bb99-5d1eb36ca4a1-ovnkube-config\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.458969 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7205a259-4e48-4108-bb99-5d1eb36ca4a1-etc-openvswitch\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.458974 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7205a259-4e48-4108-bb99-5d1eb36ca4a1-host-cni-bin\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.458936 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7205a259-4e48-4108-bb99-5d1eb36ca4a1-host-run-netns\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.459111 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/7205a259-4e48-4108-bb99-5d1eb36ca4a1-run-systemd\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.462698 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7205a259-4e48-4108-bb99-5d1eb36ca4a1-ovn-node-metrics-cert\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.494086 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8nxs\" (UniqueName: \"kubernetes.io/projected/7205a259-4e48-4108-bb99-5d1eb36ca4a1-kube-api-access-r8nxs\") pod \"ovnkube-node-l5ljn\" (UID: \"7205a259-4e48-4108-bb99-5d1eb36ca4a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:41 crc kubenswrapper[4733]: I0202 15:26:41.621970 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:42 crc kubenswrapper[4733]: I0202 15:26:42.105756 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxvkf_f5a81ada-e393-499a-aac6-7b40349958bb/ovn-acl-logging/0.log" Feb 02 15:26:42 crc kubenswrapper[4733]: I0202 15:26:42.107332 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxvkf_f5a81ada-e393-499a-aac6-7b40349958bb/ovn-controller/0.log" Feb 02 15:26:42 crc kubenswrapper[4733]: I0202 15:26:42.107973 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" event={"ID":"f5a81ada-e393-499a-aac6-7b40349958bb","Type":"ContainerDied","Data":"37078bebdb038f4f11bd31eb1708207108febd2a084514e18cd6e480f47faab9"} Feb 02 15:26:42 crc kubenswrapper[4733]: I0202 15:26:42.108022 4733 scope.go:117] "RemoveContainer" containerID="21363125c857d52144029dc46a3a1553d0b588546a6aaeab91155f3802857fc5" Feb 02 15:26:42 crc kubenswrapper[4733]: I0202 15:26:42.108080 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qxvkf" Feb 02 15:26:42 crc kubenswrapper[4733]: I0202 15:26:42.110112 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jh6vk_b76a05d6-d9c7-4183-ad90-b3131f225841/kube-multus/2.log" Feb 02 15:26:42 crc kubenswrapper[4733]: I0202 15:26:42.113499 4733 generic.go:334] "Generic (PLEG): container finished" podID="7205a259-4e48-4108-bb99-5d1eb36ca4a1" containerID="c3d1467a00399ac93748c2bd746542b5645efb1f9b23198b9994c670079c9988" exitCode=0 Feb 02 15:26:42 crc kubenswrapper[4733]: I0202 15:26:42.113546 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" event={"ID":"7205a259-4e48-4108-bb99-5d1eb36ca4a1","Type":"ContainerDied","Data":"c3d1467a00399ac93748c2bd746542b5645efb1f9b23198b9994c670079c9988"} Feb 02 15:26:42 crc kubenswrapper[4733]: I0202 15:26:42.113569 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" event={"ID":"7205a259-4e48-4108-bb99-5d1eb36ca4a1","Type":"ContainerStarted","Data":"e8ecc78d80a6c5855f7f1c3c5b029f180508899013510b90678090aab1acc3d8"} Feb 02 15:26:42 crc kubenswrapper[4733]: I0202 15:26:42.150360 4733 scope.go:117] "RemoveContainer" containerID="6efaf2216e4c1750139f080a93b4b608737dab2883de5fddfd9a99248d517b2c" Feb 02 15:26:42 crc kubenswrapper[4733]: I0202 15:26:42.179103 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qxvkf"] Feb 02 15:26:42 crc kubenswrapper[4733]: I0202 15:26:42.185872 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qxvkf"] Feb 02 15:26:42 crc kubenswrapper[4733]: I0202 15:26:42.189728 4733 scope.go:117] "RemoveContainer" containerID="1380122d1694c1286f1102e6396c6504b91f63c6eb157689dfc0ea7188c17f68" Feb 02 15:26:42 crc kubenswrapper[4733]: I0202 15:26:42.212922 4733 scope.go:117] "RemoveContainer" containerID="c5835cb53385359e9c58eaa5d4da2569c313e0fc265ad79a6c124e8cecb1a734" Feb 02 15:26:42 crc kubenswrapper[4733]: I0202 15:26:42.227350 4733 scope.go:117] "RemoveContainer" containerID="831d85055203f153381eb51d98af0cd2e4c4a0309f837545ab3d855a680133e2" Feb 02 15:26:42 crc kubenswrapper[4733]: I0202 15:26:42.242215 4733 scope.go:117] "RemoveContainer" containerID="4acf2398e2e19bd54f482cb2273c54738e61627187860bb0b6daadddc2565b63" Feb 02 15:26:42 crc kubenswrapper[4733]: I0202 15:26:42.256204 4733 scope.go:117] "RemoveContainer" containerID="ffd6dc4bf5b9fd31bfbee6d9340fe981c0c5fca03a79cef9316340ff5a8b3d0c" Feb 02 15:26:42 crc kubenswrapper[4733]: I0202 15:26:42.267365 4733 scope.go:117] "RemoveContainer" containerID="db1db75e392ebea8b387b85cabc4276d553bf2c06694cf7f294a383909f32826" Feb 02 15:26:42 crc kubenswrapper[4733]: I0202 15:26:42.280530 4733 scope.go:117] "RemoveContainer" containerID="206662e21ca79903983a16bf2082581f55e42988cdbaba89d33617277b89acbf" Feb 02 15:26:43 crc kubenswrapper[4733]: I0202 15:26:43.122016 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" event={"ID":"7205a259-4e48-4108-bb99-5d1eb36ca4a1","Type":"ContainerStarted","Data":"2b278400a2fcfd3934d9e365644cfad897dab1f2fe7549efe827e8d48c7079b7"} Feb 02 15:26:43 crc kubenswrapper[4733]: I0202 15:26:43.122407 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" event={"ID":"7205a259-4e48-4108-bb99-5d1eb36ca4a1","Type":"ContainerStarted","Data":"dd9f49f318a2ff79662286c6237d8f8d62fff53cefba0bd8eabe9ed447ea4006"} Feb 02 15:26:43 crc kubenswrapper[4733]: I0202 15:26:43.122424 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" event={"ID":"7205a259-4e48-4108-bb99-5d1eb36ca4a1","Type":"ContainerStarted","Data":"bbc3e5875b39be3a2b97301c79723299ed61dbea4f83a196a11d9d54546bd33c"} Feb 02 15:26:43 crc kubenswrapper[4733]: I0202 15:26:43.122433 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" event={"ID":"7205a259-4e48-4108-bb99-5d1eb36ca4a1","Type":"ContainerStarted","Data":"0eda60dea2cf6da7ea090c4accf1b8d8ee8c3fda96ce563249d2c63dad39a090"} Feb 02 15:26:43 crc kubenswrapper[4733]: I0202 15:26:43.122445 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" event={"ID":"7205a259-4e48-4108-bb99-5d1eb36ca4a1","Type":"ContainerStarted","Data":"f86d8bd0b1d39cd4907c8e7c86d52530fe4197ec8ec08853d6b5bc35d6bceed8"} Feb 02 15:26:43 crc kubenswrapper[4733]: I0202 15:26:43.122453 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" event={"ID":"7205a259-4e48-4108-bb99-5d1eb36ca4a1","Type":"ContainerStarted","Data":"f18dfb53fe7f19fe9011dff7081eb01f22f7dd829b4e7565e56869948fbb7cd7"} Feb 02 15:26:43 crc kubenswrapper[4733]: I0202 15:26:43.263916 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5a81ada-e393-499a-aac6-7b40349958bb" path="/var/lib/kubelet/pods/f5a81ada-e393-499a-aac6-7b40349958bb/volumes" Feb 02 15:26:46 crc kubenswrapper[4733]: I0202 15:26:46.148673 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" event={"ID":"7205a259-4e48-4108-bb99-5d1eb36ca4a1","Type":"ContainerStarted","Data":"90336eba4cff9cd3b3af5e206a04d23619b64b15d2c9d4077ea184eb0c8a3c61"} Feb 02 15:26:48 crc kubenswrapper[4733]: I0202 15:26:48.168456 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" event={"ID":"7205a259-4e48-4108-bb99-5d1eb36ca4a1","Type":"ContainerStarted","Data":"ae2495e5267c8fe0f4540fd67c16083ee9b57d3ac3f4f34807f0469b9663e76c"} Feb 02 15:26:48 crc kubenswrapper[4733]: I0202 15:26:48.169073 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:48 crc kubenswrapper[4733]: I0202 15:26:48.169136 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:48 crc kubenswrapper[4733]: I0202 15:26:48.202697 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:48 crc kubenswrapper[4733]: I0202 15:26:48.208067 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" podStartSLOduration=7.20804989 podStartE2EDuration="7.20804989s" podCreationTimestamp="2026-02-02 15:26:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:26:48.204875805 +0000 UTC m=+691.656337183" watchObservedRunningTime="2026-02-02 15:26:48.20804989 +0000 UTC m=+691.659511248" Feb 02 15:26:49 crc kubenswrapper[4733]: I0202 15:26:49.174265 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:49 crc kubenswrapper[4733]: I0202 15:26:49.215963 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:26:56 crc kubenswrapper[4733]: I0202 15:26:56.255569 4733 scope.go:117] "RemoveContainer" containerID="0746ffd4e82c3554d4f63c139295a9cedb77073db5dda7ba15e32d519c0ed868" Feb 02 15:26:56 crc kubenswrapper[4733]: E0202 15:26:56.256477 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-jh6vk_openshift-multus(b76a05d6-d9c7-4183-ad90-b3131f225841)\"" pod="openshift-multus/multus-jh6vk" podUID="b76a05d6-d9c7-4183-ad90-b3131f225841" Feb 02 15:27:05 crc kubenswrapper[4733]: I0202 15:27:05.740899 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg"] Feb 02 15:27:05 crc kubenswrapper[4733]: I0202 15:27:05.743141 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg" Feb 02 15:27:05 crc kubenswrapper[4733]: I0202 15:27:05.748025 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Feb 02 15:27:05 crc kubenswrapper[4733]: I0202 15:27:05.779130 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg"] Feb 02 15:27:05 crc kubenswrapper[4733]: I0202 15:27:05.804377 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e8c251f4-787d-49f1-8ec3-32f6c5dff2bb-bundle\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg\" (UID: \"e8c251f4-787d-49f1-8ec3-32f6c5dff2bb\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg" Feb 02 15:27:05 crc kubenswrapper[4733]: I0202 15:27:05.804432 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phnqx\" (UniqueName: \"kubernetes.io/projected/e8c251f4-787d-49f1-8ec3-32f6c5dff2bb-kube-api-access-phnqx\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg\" (UID: \"e8c251f4-787d-49f1-8ec3-32f6c5dff2bb\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg" Feb 02 15:27:05 crc kubenswrapper[4733]: I0202 15:27:05.804603 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e8c251f4-787d-49f1-8ec3-32f6c5dff2bb-util\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg\" (UID: \"e8c251f4-787d-49f1-8ec3-32f6c5dff2bb\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg" Feb 02 15:27:05 crc kubenswrapper[4733]: I0202 15:27:05.905848 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e8c251f4-787d-49f1-8ec3-32f6c5dff2bb-util\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg\" (UID: \"e8c251f4-787d-49f1-8ec3-32f6c5dff2bb\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg" Feb 02 15:27:05 crc kubenswrapper[4733]: I0202 15:27:05.906022 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e8c251f4-787d-49f1-8ec3-32f6c5dff2bb-bundle\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg\" (UID: \"e8c251f4-787d-49f1-8ec3-32f6c5dff2bb\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg" Feb 02 15:27:05 crc kubenswrapper[4733]: I0202 15:27:05.906084 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phnqx\" (UniqueName: \"kubernetes.io/projected/e8c251f4-787d-49f1-8ec3-32f6c5dff2bb-kube-api-access-phnqx\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg\" (UID: \"e8c251f4-787d-49f1-8ec3-32f6c5dff2bb\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg" Feb 02 15:27:05 crc kubenswrapper[4733]: I0202 15:27:05.906629 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e8c251f4-787d-49f1-8ec3-32f6c5dff2bb-util\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg\" (UID: \"e8c251f4-787d-49f1-8ec3-32f6c5dff2bb\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg" Feb 02 15:27:05 crc kubenswrapper[4733]: I0202 15:27:05.906751 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e8c251f4-787d-49f1-8ec3-32f6c5dff2bb-bundle\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg\" (UID: \"e8c251f4-787d-49f1-8ec3-32f6c5dff2bb\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg" Feb 02 15:27:05 crc kubenswrapper[4733]: I0202 15:27:05.946569 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phnqx\" (UniqueName: \"kubernetes.io/projected/e8c251f4-787d-49f1-8ec3-32f6c5dff2bb-kube-api-access-phnqx\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg\" (UID: \"e8c251f4-787d-49f1-8ec3-32f6c5dff2bb\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg" Feb 02 15:27:06 crc kubenswrapper[4733]: I0202 15:27:06.071324 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg" Feb 02 15:27:06 crc kubenswrapper[4733]: E0202 15:27:06.107372 4733 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg_openshift-marketplace_e8c251f4-787d-49f1-8ec3-32f6c5dff2bb_0(c6efbe363883df43e5d3419b5afa3aebcc3d5d4a8c32daef9f9832834ed43da8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 02 15:27:06 crc kubenswrapper[4733]: E0202 15:27:06.107446 4733 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg_openshift-marketplace_e8c251f4-787d-49f1-8ec3-32f6c5dff2bb_0(c6efbe363883df43e5d3419b5afa3aebcc3d5d4a8c32daef9f9832834ed43da8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg" Feb 02 15:27:06 crc kubenswrapper[4733]: E0202 15:27:06.107473 4733 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg_openshift-marketplace_e8c251f4-787d-49f1-8ec3-32f6c5dff2bb_0(c6efbe363883df43e5d3419b5afa3aebcc3d5d4a8c32daef9f9832834ed43da8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg" Feb 02 15:27:06 crc kubenswrapper[4733]: E0202 15:27:06.107520 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg_openshift-marketplace(e8c251f4-787d-49f1-8ec3-32f6c5dff2bb)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg_openshift-marketplace(e8c251f4-787d-49f1-8ec3-32f6c5dff2bb)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg_openshift-marketplace_e8c251f4-787d-49f1-8ec3-32f6c5dff2bb_0(c6efbe363883df43e5d3419b5afa3aebcc3d5d4a8c32daef9f9832834ed43da8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg" podUID="e8c251f4-787d-49f1-8ec3-32f6c5dff2bb" Feb 02 15:27:06 crc kubenswrapper[4733]: I0202 15:27:06.295997 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg" Feb 02 15:27:06 crc kubenswrapper[4733]: I0202 15:27:06.297488 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg" Feb 02 15:27:06 crc kubenswrapper[4733]: E0202 15:27:06.340215 4733 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg_openshift-marketplace_e8c251f4-787d-49f1-8ec3-32f6c5dff2bb_0(95d22ea620e90ccff2ce2bb43ea1c70e5b93daa20e544a4e53352d6fea160efd): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 02 15:27:06 crc kubenswrapper[4733]: E0202 15:27:06.340317 4733 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg_openshift-marketplace_e8c251f4-787d-49f1-8ec3-32f6c5dff2bb_0(95d22ea620e90ccff2ce2bb43ea1c70e5b93daa20e544a4e53352d6fea160efd): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg" Feb 02 15:27:06 crc kubenswrapper[4733]: E0202 15:27:06.340373 4733 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg_openshift-marketplace_e8c251f4-787d-49f1-8ec3-32f6c5dff2bb_0(95d22ea620e90ccff2ce2bb43ea1c70e5b93daa20e544a4e53352d6fea160efd): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg" Feb 02 15:27:06 crc kubenswrapper[4733]: E0202 15:27:06.340468 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg_openshift-marketplace(e8c251f4-787d-49f1-8ec3-32f6c5dff2bb)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg_openshift-marketplace(e8c251f4-787d-49f1-8ec3-32f6c5dff2bb)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg_openshift-marketplace_e8c251f4-787d-49f1-8ec3-32f6c5dff2bb_0(95d22ea620e90ccff2ce2bb43ea1c70e5b93daa20e544a4e53352d6fea160efd): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg" podUID="e8c251f4-787d-49f1-8ec3-32f6c5dff2bb" Feb 02 15:27:08 crc kubenswrapper[4733]: I0202 15:27:08.255470 4733 scope.go:117] "RemoveContainer" containerID="0746ffd4e82c3554d4f63c139295a9cedb77073db5dda7ba15e32d519c0ed868" Feb 02 15:27:09 crc kubenswrapper[4733]: I0202 15:27:09.318441 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jh6vk_b76a05d6-d9c7-4183-ad90-b3131f225841/kube-multus/2.log" Feb 02 15:27:09 crc kubenswrapper[4733]: I0202 15:27:09.318844 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jh6vk" event={"ID":"b76a05d6-d9c7-4183-ad90-b3131f225841","Type":"ContainerStarted","Data":"fccff0d3f14082d5d6c3b40bd077530cde96d6b0d793ce51385ce1a28d5b64d1"} Feb 02 15:27:11 crc kubenswrapper[4733]: I0202 15:27:11.643182 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-l5ljn" Feb 02 15:27:17 crc kubenswrapper[4733]: I0202 15:27:17.254743 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg" Feb 02 15:27:17 crc kubenswrapper[4733]: I0202 15:27:17.260994 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg" Feb 02 15:27:17 crc kubenswrapper[4733]: I0202 15:27:17.522364 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg"] Feb 02 15:27:17 crc kubenswrapper[4733]: W0202 15:27:17.527574 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode8c251f4_787d_49f1_8ec3_32f6c5dff2bb.slice/crio-f51a38066cfcfde31977a08de2127de1349dc822490a81601c5429a8a59bb600 WatchSource:0}: Error finding container f51a38066cfcfde31977a08de2127de1349dc822490a81601c5429a8a59bb600: Status 404 returned error can't find the container with id f51a38066cfcfde31977a08de2127de1349dc822490a81601c5429a8a59bb600 Feb 02 15:27:18 crc kubenswrapper[4733]: I0202 15:27:18.381006 4733 generic.go:334] "Generic (PLEG): container finished" podID="e8c251f4-787d-49f1-8ec3-32f6c5dff2bb" containerID="67b4022a4cc3fa2f0ffc363baae746ccc58b8f048024beafaa8bb4096c74ef6a" exitCode=0 Feb 02 15:27:18 crc kubenswrapper[4733]: I0202 15:27:18.381143 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg" event={"ID":"e8c251f4-787d-49f1-8ec3-32f6c5dff2bb","Type":"ContainerDied","Data":"67b4022a4cc3fa2f0ffc363baae746ccc58b8f048024beafaa8bb4096c74ef6a"} Feb 02 15:27:18 crc kubenswrapper[4733]: I0202 15:27:18.381563 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg" event={"ID":"e8c251f4-787d-49f1-8ec3-32f6c5dff2bb","Type":"ContainerStarted","Data":"f51a38066cfcfde31977a08de2127de1349dc822490a81601c5429a8a59bb600"} Feb 02 15:27:18 crc kubenswrapper[4733]: I0202 15:27:18.385048 4733 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 02 15:27:21 crc kubenswrapper[4733]: I0202 15:27:21.408580 4733 generic.go:334] "Generic (PLEG): container finished" podID="e8c251f4-787d-49f1-8ec3-32f6c5dff2bb" containerID="10f470f677bd64dfd10e5c7e34f6e89f825683fc47e3936331ea13d69ede0cf6" exitCode=0 Feb 02 15:27:21 crc kubenswrapper[4733]: I0202 15:27:21.408765 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg" event={"ID":"e8c251f4-787d-49f1-8ec3-32f6c5dff2bb","Type":"ContainerDied","Data":"10f470f677bd64dfd10e5c7e34f6e89f825683fc47e3936331ea13d69ede0cf6"} Feb 02 15:27:22 crc kubenswrapper[4733]: I0202 15:27:22.420919 4733 generic.go:334] "Generic (PLEG): container finished" podID="e8c251f4-787d-49f1-8ec3-32f6c5dff2bb" containerID="af3b14e63653d4cc6d973881dca34c05abb40324d37a94c2f24ec20f3fc1f32d" exitCode=0 Feb 02 15:27:22 crc kubenswrapper[4733]: I0202 15:27:22.421067 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg" event={"ID":"e8c251f4-787d-49f1-8ec3-32f6c5dff2bb","Type":"ContainerDied","Data":"af3b14e63653d4cc6d973881dca34c05abb40324d37a94c2f24ec20f3fc1f32d"} Feb 02 15:27:23 crc kubenswrapper[4733]: I0202 15:27:23.690244 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg" Feb 02 15:27:23 crc kubenswrapper[4733]: I0202 15:27:23.854085 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-phnqx\" (UniqueName: \"kubernetes.io/projected/e8c251f4-787d-49f1-8ec3-32f6c5dff2bb-kube-api-access-phnqx\") pod \"e8c251f4-787d-49f1-8ec3-32f6c5dff2bb\" (UID: \"e8c251f4-787d-49f1-8ec3-32f6c5dff2bb\") " Feb 02 15:27:23 crc kubenswrapper[4733]: I0202 15:27:23.854272 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e8c251f4-787d-49f1-8ec3-32f6c5dff2bb-bundle\") pod \"e8c251f4-787d-49f1-8ec3-32f6c5dff2bb\" (UID: \"e8c251f4-787d-49f1-8ec3-32f6c5dff2bb\") " Feb 02 15:27:23 crc kubenswrapper[4733]: I0202 15:27:23.854378 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e8c251f4-787d-49f1-8ec3-32f6c5dff2bb-util\") pod \"e8c251f4-787d-49f1-8ec3-32f6c5dff2bb\" (UID: \"e8c251f4-787d-49f1-8ec3-32f6c5dff2bb\") " Feb 02 15:27:23 crc kubenswrapper[4733]: I0202 15:27:23.856112 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8c251f4-787d-49f1-8ec3-32f6c5dff2bb-bundle" (OuterVolumeSpecName: "bundle") pod "e8c251f4-787d-49f1-8ec3-32f6c5dff2bb" (UID: "e8c251f4-787d-49f1-8ec3-32f6c5dff2bb"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:27:23 crc kubenswrapper[4733]: I0202 15:27:23.863505 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8c251f4-787d-49f1-8ec3-32f6c5dff2bb-kube-api-access-phnqx" (OuterVolumeSpecName: "kube-api-access-phnqx") pod "e8c251f4-787d-49f1-8ec3-32f6c5dff2bb" (UID: "e8c251f4-787d-49f1-8ec3-32f6c5dff2bb"). InnerVolumeSpecName "kube-api-access-phnqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:27:23 crc kubenswrapper[4733]: I0202 15:27:23.876850 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8c251f4-787d-49f1-8ec3-32f6c5dff2bb-util" (OuterVolumeSpecName: "util") pod "e8c251f4-787d-49f1-8ec3-32f6c5dff2bb" (UID: "e8c251f4-787d-49f1-8ec3-32f6c5dff2bb"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:27:23 crc kubenswrapper[4733]: I0202 15:27:23.956325 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-phnqx\" (UniqueName: \"kubernetes.io/projected/e8c251f4-787d-49f1-8ec3-32f6c5dff2bb-kube-api-access-phnqx\") on node \"crc\" DevicePath \"\"" Feb 02 15:27:23 crc kubenswrapper[4733]: I0202 15:27:23.956368 4733 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e8c251f4-787d-49f1-8ec3-32f6c5dff2bb-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 15:27:23 crc kubenswrapper[4733]: I0202 15:27:23.956380 4733 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e8c251f4-787d-49f1-8ec3-32f6c5dff2bb-util\") on node \"crc\" DevicePath \"\"" Feb 02 15:27:24 crc kubenswrapper[4733]: I0202 15:27:24.438494 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg" event={"ID":"e8c251f4-787d-49f1-8ec3-32f6c5dff2bb","Type":"ContainerDied","Data":"f51a38066cfcfde31977a08de2127de1349dc822490a81601c5429a8a59bb600"} Feb 02 15:27:24 crc kubenswrapper[4733]: I0202 15:27:24.438566 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f51a38066cfcfde31977a08de2127de1349dc822490a81601c5429a8a59bb600" Feb 02 15:27:24 crc kubenswrapper[4733]: I0202 15:27:24.438578 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg" Feb 02 15:27:34 crc kubenswrapper[4733]: I0202 15:27:34.018763 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-54b6695bd-mwgx2"] Feb 02 15:27:34 crc kubenswrapper[4733]: E0202 15:27:34.019571 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8c251f4-787d-49f1-8ec3-32f6c5dff2bb" containerName="util" Feb 02 15:27:34 crc kubenswrapper[4733]: I0202 15:27:34.019587 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8c251f4-787d-49f1-8ec3-32f6c5dff2bb" containerName="util" Feb 02 15:27:34 crc kubenswrapper[4733]: E0202 15:27:34.019614 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8c251f4-787d-49f1-8ec3-32f6c5dff2bb" containerName="extract" Feb 02 15:27:34 crc kubenswrapper[4733]: I0202 15:27:34.019623 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8c251f4-787d-49f1-8ec3-32f6c5dff2bb" containerName="extract" Feb 02 15:27:34 crc kubenswrapper[4733]: E0202 15:27:34.019631 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8c251f4-787d-49f1-8ec3-32f6c5dff2bb" containerName="pull" Feb 02 15:27:34 crc kubenswrapper[4733]: I0202 15:27:34.019638 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8c251f4-787d-49f1-8ec3-32f6c5dff2bb" containerName="pull" Feb 02 15:27:34 crc kubenswrapper[4733]: I0202 15:27:34.019745 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8c251f4-787d-49f1-8ec3-32f6c5dff2bb" containerName="extract" Feb 02 15:27:34 crc kubenswrapper[4733]: I0202 15:27:34.020246 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-54b6695bd-mwgx2" Feb 02 15:27:34 crc kubenswrapper[4733]: I0202 15:27:34.022370 4733 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-svqkq" Feb 02 15:27:34 crc kubenswrapper[4733]: I0202 15:27:34.024275 4733 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Feb 02 15:27:34 crc kubenswrapper[4733]: I0202 15:27:34.024310 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Feb 02 15:27:34 crc kubenswrapper[4733]: I0202 15:27:34.024407 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Feb 02 15:27:34 crc kubenswrapper[4733]: I0202 15:27:34.024674 4733 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Feb 02 15:27:34 crc kubenswrapper[4733]: I0202 15:27:34.039290 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-54b6695bd-mwgx2"] Feb 02 15:27:34 crc kubenswrapper[4733]: I0202 15:27:34.073006 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2e7a0012-8704-4613-9643-ccb780c9d76d-webhook-cert\") pod \"metallb-operator-controller-manager-54b6695bd-mwgx2\" (UID: \"2e7a0012-8704-4613-9643-ccb780c9d76d\") " pod="metallb-system/metallb-operator-controller-manager-54b6695bd-mwgx2" Feb 02 15:27:34 crc kubenswrapper[4733]: I0202 15:27:34.073065 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2e7a0012-8704-4613-9643-ccb780c9d76d-apiservice-cert\") pod \"metallb-operator-controller-manager-54b6695bd-mwgx2\" (UID: \"2e7a0012-8704-4613-9643-ccb780c9d76d\") " pod="metallb-system/metallb-operator-controller-manager-54b6695bd-mwgx2" Feb 02 15:27:34 crc kubenswrapper[4733]: I0202 15:27:34.073124 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66m8c\" (UniqueName: \"kubernetes.io/projected/2e7a0012-8704-4613-9643-ccb780c9d76d-kube-api-access-66m8c\") pod \"metallb-operator-controller-manager-54b6695bd-mwgx2\" (UID: \"2e7a0012-8704-4613-9643-ccb780c9d76d\") " pod="metallb-system/metallb-operator-controller-manager-54b6695bd-mwgx2" Feb 02 15:27:34 crc kubenswrapper[4733]: I0202 15:27:34.174140 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2e7a0012-8704-4613-9643-ccb780c9d76d-webhook-cert\") pod \"metallb-operator-controller-manager-54b6695bd-mwgx2\" (UID: \"2e7a0012-8704-4613-9643-ccb780c9d76d\") " pod="metallb-system/metallb-operator-controller-manager-54b6695bd-mwgx2" Feb 02 15:27:34 crc kubenswrapper[4733]: I0202 15:27:34.174209 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2e7a0012-8704-4613-9643-ccb780c9d76d-apiservice-cert\") pod \"metallb-operator-controller-manager-54b6695bd-mwgx2\" (UID: \"2e7a0012-8704-4613-9643-ccb780c9d76d\") " pod="metallb-system/metallb-operator-controller-manager-54b6695bd-mwgx2" Feb 02 15:27:34 crc kubenswrapper[4733]: I0202 15:27:34.174259 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66m8c\" (UniqueName: \"kubernetes.io/projected/2e7a0012-8704-4613-9643-ccb780c9d76d-kube-api-access-66m8c\") pod \"metallb-operator-controller-manager-54b6695bd-mwgx2\" (UID: \"2e7a0012-8704-4613-9643-ccb780c9d76d\") " pod="metallb-system/metallb-operator-controller-manager-54b6695bd-mwgx2" Feb 02 15:27:34 crc kubenswrapper[4733]: I0202 15:27:34.181315 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2e7a0012-8704-4613-9643-ccb780c9d76d-apiservice-cert\") pod \"metallb-operator-controller-manager-54b6695bd-mwgx2\" (UID: \"2e7a0012-8704-4613-9643-ccb780c9d76d\") " pod="metallb-system/metallb-operator-controller-manager-54b6695bd-mwgx2" Feb 02 15:27:34 crc kubenswrapper[4733]: I0202 15:27:34.186225 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2e7a0012-8704-4613-9643-ccb780c9d76d-webhook-cert\") pod \"metallb-operator-controller-manager-54b6695bd-mwgx2\" (UID: \"2e7a0012-8704-4613-9643-ccb780c9d76d\") " pod="metallb-system/metallb-operator-controller-manager-54b6695bd-mwgx2" Feb 02 15:27:34 crc kubenswrapper[4733]: I0202 15:27:34.190699 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66m8c\" (UniqueName: \"kubernetes.io/projected/2e7a0012-8704-4613-9643-ccb780c9d76d-kube-api-access-66m8c\") pod \"metallb-operator-controller-manager-54b6695bd-mwgx2\" (UID: \"2e7a0012-8704-4613-9643-ccb780c9d76d\") " pod="metallb-system/metallb-operator-controller-manager-54b6695bd-mwgx2" Feb 02 15:27:34 crc kubenswrapper[4733]: I0202 15:27:34.251888 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-66bb599b85-wt9zk"] Feb 02 15:27:34 crc kubenswrapper[4733]: I0202 15:27:34.252549 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-66bb599b85-wt9zk" Feb 02 15:27:34 crc kubenswrapper[4733]: I0202 15:27:34.254821 4733 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Feb 02 15:27:34 crc kubenswrapper[4733]: I0202 15:27:34.255138 4733 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-zm49s" Feb 02 15:27:34 crc kubenswrapper[4733]: I0202 15:27:34.255723 4733 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Feb 02 15:27:34 crc kubenswrapper[4733]: I0202 15:27:34.265151 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-66bb599b85-wt9zk"] Feb 02 15:27:34 crc kubenswrapper[4733]: I0202 15:27:34.336903 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-54b6695bd-mwgx2" Feb 02 15:27:34 crc kubenswrapper[4733]: I0202 15:27:34.377813 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ffb4086a-a662-4380-8dd0-c218d36e7c31-apiservice-cert\") pod \"metallb-operator-webhook-server-66bb599b85-wt9zk\" (UID: \"ffb4086a-a662-4380-8dd0-c218d36e7c31\") " pod="metallb-system/metallb-operator-webhook-server-66bb599b85-wt9zk" Feb 02 15:27:34 crc kubenswrapper[4733]: I0202 15:27:34.377891 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccns8\" (UniqueName: \"kubernetes.io/projected/ffb4086a-a662-4380-8dd0-c218d36e7c31-kube-api-access-ccns8\") pod \"metallb-operator-webhook-server-66bb599b85-wt9zk\" (UID: \"ffb4086a-a662-4380-8dd0-c218d36e7c31\") " pod="metallb-system/metallb-operator-webhook-server-66bb599b85-wt9zk" Feb 02 15:27:34 crc kubenswrapper[4733]: I0202 15:27:34.377936 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ffb4086a-a662-4380-8dd0-c218d36e7c31-webhook-cert\") pod \"metallb-operator-webhook-server-66bb599b85-wt9zk\" (UID: \"ffb4086a-a662-4380-8dd0-c218d36e7c31\") " pod="metallb-system/metallb-operator-webhook-server-66bb599b85-wt9zk" Feb 02 15:27:34 crc kubenswrapper[4733]: I0202 15:27:34.479630 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ffb4086a-a662-4380-8dd0-c218d36e7c31-apiservice-cert\") pod \"metallb-operator-webhook-server-66bb599b85-wt9zk\" (UID: \"ffb4086a-a662-4380-8dd0-c218d36e7c31\") " pod="metallb-system/metallb-operator-webhook-server-66bb599b85-wt9zk" Feb 02 15:27:34 crc kubenswrapper[4733]: I0202 15:27:34.479934 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccns8\" (UniqueName: \"kubernetes.io/projected/ffb4086a-a662-4380-8dd0-c218d36e7c31-kube-api-access-ccns8\") pod \"metallb-operator-webhook-server-66bb599b85-wt9zk\" (UID: \"ffb4086a-a662-4380-8dd0-c218d36e7c31\") " pod="metallb-system/metallb-operator-webhook-server-66bb599b85-wt9zk" Feb 02 15:27:34 crc kubenswrapper[4733]: I0202 15:27:34.479975 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ffb4086a-a662-4380-8dd0-c218d36e7c31-webhook-cert\") pod \"metallb-operator-webhook-server-66bb599b85-wt9zk\" (UID: \"ffb4086a-a662-4380-8dd0-c218d36e7c31\") " pod="metallb-system/metallb-operator-webhook-server-66bb599b85-wt9zk" Feb 02 15:27:34 crc kubenswrapper[4733]: I0202 15:27:34.506003 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ffb4086a-a662-4380-8dd0-c218d36e7c31-apiservice-cert\") pod \"metallb-operator-webhook-server-66bb599b85-wt9zk\" (UID: \"ffb4086a-a662-4380-8dd0-c218d36e7c31\") " pod="metallb-system/metallb-operator-webhook-server-66bb599b85-wt9zk" Feb 02 15:27:34 crc kubenswrapper[4733]: I0202 15:27:34.508210 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccns8\" (UniqueName: \"kubernetes.io/projected/ffb4086a-a662-4380-8dd0-c218d36e7c31-kube-api-access-ccns8\") pod \"metallb-operator-webhook-server-66bb599b85-wt9zk\" (UID: \"ffb4086a-a662-4380-8dd0-c218d36e7c31\") " pod="metallb-system/metallb-operator-webhook-server-66bb599b85-wt9zk" Feb 02 15:27:34 crc kubenswrapper[4733]: I0202 15:27:34.512731 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ffb4086a-a662-4380-8dd0-c218d36e7c31-webhook-cert\") pod \"metallb-operator-webhook-server-66bb599b85-wt9zk\" (UID: \"ffb4086a-a662-4380-8dd0-c218d36e7c31\") " pod="metallb-system/metallb-operator-webhook-server-66bb599b85-wt9zk" Feb 02 15:27:34 crc kubenswrapper[4733]: I0202 15:27:34.563489 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-54b6695bd-mwgx2"] Feb 02 15:27:34 crc kubenswrapper[4733]: W0202 15:27:34.570770 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2e7a0012_8704_4613_9643_ccb780c9d76d.slice/crio-e139129782f7e77100ba7d2987fb159f5d577d3e1535b9a1475bf6e0f09c3139 WatchSource:0}: Error finding container e139129782f7e77100ba7d2987fb159f5d577d3e1535b9a1475bf6e0f09c3139: Status 404 returned error can't find the container with id e139129782f7e77100ba7d2987fb159f5d577d3e1535b9a1475bf6e0f09c3139 Feb 02 15:27:34 crc kubenswrapper[4733]: I0202 15:27:34.571698 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-66bb599b85-wt9zk" Feb 02 15:27:35 crc kubenswrapper[4733]: I0202 15:27:34.999873 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-66bb599b85-wt9zk"] Feb 02 15:27:35 crc kubenswrapper[4733]: W0202 15:27:35.008695 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podffb4086a_a662_4380_8dd0_c218d36e7c31.slice/crio-3a9950a323c9ddc4eccc8cbd29951676636cfaeb201d6205407b71992df4b6b5 WatchSource:0}: Error finding container 3a9950a323c9ddc4eccc8cbd29951676636cfaeb201d6205407b71992df4b6b5: Status 404 returned error can't find the container with id 3a9950a323c9ddc4eccc8cbd29951676636cfaeb201d6205407b71992df4b6b5 Feb 02 15:27:35 crc kubenswrapper[4733]: I0202 15:27:35.498609 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-54b6695bd-mwgx2" event={"ID":"2e7a0012-8704-4613-9643-ccb780c9d76d","Type":"ContainerStarted","Data":"e139129782f7e77100ba7d2987fb159f5d577d3e1535b9a1475bf6e0f09c3139"} Feb 02 15:27:35 crc kubenswrapper[4733]: I0202 15:27:35.501261 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-66bb599b85-wt9zk" event={"ID":"ffb4086a-a662-4380-8dd0-c218d36e7c31","Type":"ContainerStarted","Data":"3a9950a323c9ddc4eccc8cbd29951676636cfaeb201d6205407b71992df4b6b5"} Feb 02 15:27:40 crc kubenswrapper[4733]: I0202 15:27:40.544852 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-66bb599b85-wt9zk" event={"ID":"ffb4086a-a662-4380-8dd0-c218d36e7c31","Type":"ContainerStarted","Data":"e595ff6df02f8ca4db86f9c3a33e63b24980eff65287f7c135bd46b24c3e5262"} Feb 02 15:27:40 crc kubenswrapper[4733]: I0202 15:27:40.545478 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-66bb599b85-wt9zk" Feb 02 15:27:40 crc kubenswrapper[4733]: I0202 15:27:40.547629 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-54b6695bd-mwgx2" event={"ID":"2e7a0012-8704-4613-9643-ccb780c9d76d","Type":"ContainerStarted","Data":"bada2ac03fad88d65e33d140cec5996623f262d3046900ac50685a741764134f"} Feb 02 15:27:40 crc kubenswrapper[4733]: I0202 15:27:40.547893 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-54b6695bd-mwgx2" Feb 02 15:27:40 crc kubenswrapper[4733]: I0202 15:27:40.576257 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-66bb599b85-wt9zk" podStartSLOduration=1.62866789 podStartE2EDuration="6.576231147s" podCreationTimestamp="2026-02-02 15:27:34 +0000 UTC" firstStartedPulling="2026-02-02 15:27:35.010696094 +0000 UTC m=+738.462157442" lastFinishedPulling="2026-02-02 15:27:39.958259341 +0000 UTC m=+743.409720699" observedRunningTime="2026-02-02 15:27:40.571363406 +0000 UTC m=+744.022824754" watchObservedRunningTime="2026-02-02 15:27:40.576231147 +0000 UTC m=+744.027692515" Feb 02 15:27:40 crc kubenswrapper[4733]: I0202 15:27:40.596633 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-54b6695bd-mwgx2" podStartSLOduration=1.228067369 podStartE2EDuration="6.596616724s" podCreationTimestamp="2026-02-02 15:27:34 +0000 UTC" firstStartedPulling="2026-02-02 15:27:34.572551605 +0000 UTC m=+738.024012963" lastFinishedPulling="2026-02-02 15:27:39.94110096 +0000 UTC m=+743.392562318" observedRunningTime="2026-02-02 15:27:40.594963099 +0000 UTC m=+744.046424457" watchObservedRunningTime="2026-02-02 15:27:40.596616724 +0000 UTC m=+744.048078082" Feb 02 15:27:54 crc kubenswrapper[4733]: I0202 15:27:54.219555 4733 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 02 15:27:54 crc kubenswrapper[4733]: I0202 15:27:54.576549 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-66bb599b85-wt9zk" Feb 02 15:28:00 crc kubenswrapper[4733]: I0202 15:28:00.838585 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9gnj4"] Feb 02 15:28:00 crc kubenswrapper[4733]: I0202 15:28:00.840817 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9gnj4" Feb 02 15:28:00 crc kubenswrapper[4733]: I0202 15:28:00.853811 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9gnj4"] Feb 02 15:28:01 crc kubenswrapper[4733]: I0202 15:28:01.023179 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58vdc\" (UniqueName: \"kubernetes.io/projected/501a8c1d-5254-4aac-b3c5-94a56b9e0fc0-kube-api-access-58vdc\") pod \"redhat-operators-9gnj4\" (UID: \"501a8c1d-5254-4aac-b3c5-94a56b9e0fc0\") " pod="openshift-marketplace/redhat-operators-9gnj4" Feb 02 15:28:01 crc kubenswrapper[4733]: I0202 15:28:01.023238 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/501a8c1d-5254-4aac-b3c5-94a56b9e0fc0-utilities\") pod \"redhat-operators-9gnj4\" (UID: \"501a8c1d-5254-4aac-b3c5-94a56b9e0fc0\") " pod="openshift-marketplace/redhat-operators-9gnj4" Feb 02 15:28:01 crc kubenswrapper[4733]: I0202 15:28:01.023283 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/501a8c1d-5254-4aac-b3c5-94a56b9e0fc0-catalog-content\") pod \"redhat-operators-9gnj4\" (UID: \"501a8c1d-5254-4aac-b3c5-94a56b9e0fc0\") " pod="openshift-marketplace/redhat-operators-9gnj4" Feb 02 15:28:01 crc kubenswrapper[4733]: I0202 15:28:01.124153 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/501a8c1d-5254-4aac-b3c5-94a56b9e0fc0-utilities\") pod \"redhat-operators-9gnj4\" (UID: \"501a8c1d-5254-4aac-b3c5-94a56b9e0fc0\") " pod="openshift-marketplace/redhat-operators-9gnj4" Feb 02 15:28:01 crc kubenswrapper[4733]: I0202 15:28:01.124236 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/501a8c1d-5254-4aac-b3c5-94a56b9e0fc0-catalog-content\") pod \"redhat-operators-9gnj4\" (UID: \"501a8c1d-5254-4aac-b3c5-94a56b9e0fc0\") " pod="openshift-marketplace/redhat-operators-9gnj4" Feb 02 15:28:01 crc kubenswrapper[4733]: I0202 15:28:01.124275 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58vdc\" (UniqueName: \"kubernetes.io/projected/501a8c1d-5254-4aac-b3c5-94a56b9e0fc0-kube-api-access-58vdc\") pod \"redhat-operators-9gnj4\" (UID: \"501a8c1d-5254-4aac-b3c5-94a56b9e0fc0\") " pod="openshift-marketplace/redhat-operators-9gnj4" Feb 02 15:28:01 crc kubenswrapper[4733]: I0202 15:28:01.124957 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/501a8c1d-5254-4aac-b3c5-94a56b9e0fc0-utilities\") pod \"redhat-operators-9gnj4\" (UID: \"501a8c1d-5254-4aac-b3c5-94a56b9e0fc0\") " pod="openshift-marketplace/redhat-operators-9gnj4" Feb 02 15:28:01 crc kubenswrapper[4733]: I0202 15:28:01.125104 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/501a8c1d-5254-4aac-b3c5-94a56b9e0fc0-catalog-content\") pod \"redhat-operators-9gnj4\" (UID: \"501a8c1d-5254-4aac-b3c5-94a56b9e0fc0\") " pod="openshift-marketplace/redhat-operators-9gnj4" Feb 02 15:28:01 crc kubenswrapper[4733]: I0202 15:28:01.148431 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58vdc\" (UniqueName: \"kubernetes.io/projected/501a8c1d-5254-4aac-b3c5-94a56b9e0fc0-kube-api-access-58vdc\") pod \"redhat-operators-9gnj4\" (UID: \"501a8c1d-5254-4aac-b3c5-94a56b9e0fc0\") " pod="openshift-marketplace/redhat-operators-9gnj4" Feb 02 15:28:01 crc kubenswrapper[4733]: I0202 15:28:01.165639 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9gnj4" Feb 02 15:28:01 crc kubenswrapper[4733]: I0202 15:28:01.359098 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9gnj4"] Feb 02 15:28:01 crc kubenswrapper[4733]: I0202 15:28:01.668127 4733 generic.go:334] "Generic (PLEG): container finished" podID="501a8c1d-5254-4aac-b3c5-94a56b9e0fc0" containerID="c8901e6c33323ad3b5bdbe10291cb7d14bed6ddaee9abd04ef234f19fa80684d" exitCode=0 Feb 02 15:28:01 crc kubenswrapper[4733]: I0202 15:28:01.668299 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9gnj4" event={"ID":"501a8c1d-5254-4aac-b3c5-94a56b9e0fc0","Type":"ContainerDied","Data":"c8901e6c33323ad3b5bdbe10291cb7d14bed6ddaee9abd04ef234f19fa80684d"} Feb 02 15:28:01 crc kubenswrapper[4733]: I0202 15:28:01.668484 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9gnj4" event={"ID":"501a8c1d-5254-4aac-b3c5-94a56b9e0fc0","Type":"ContainerStarted","Data":"ea087ab9d4010689b38db8c58c6f1fd8316053ff6d447d24f390ff9d36fefd07"} Feb 02 15:28:03 crc kubenswrapper[4733]: I0202 15:28:03.683514 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9gnj4" event={"ID":"501a8c1d-5254-4aac-b3c5-94a56b9e0fc0","Type":"ContainerStarted","Data":"3c65c98c0c190da476bcee6ca1e2c1bc0b5f9be0c1dc50cc9bcf405e5dd47c68"} Feb 02 15:28:04 crc kubenswrapper[4733]: I0202 15:28:04.691200 4733 generic.go:334] "Generic (PLEG): container finished" podID="501a8c1d-5254-4aac-b3c5-94a56b9e0fc0" containerID="3c65c98c0c190da476bcee6ca1e2c1bc0b5f9be0c1dc50cc9bcf405e5dd47c68" exitCode=0 Feb 02 15:28:04 crc kubenswrapper[4733]: I0202 15:28:04.691272 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9gnj4" event={"ID":"501a8c1d-5254-4aac-b3c5-94a56b9e0fc0","Type":"ContainerDied","Data":"3c65c98c0c190da476bcee6ca1e2c1bc0b5f9be0c1dc50cc9bcf405e5dd47c68"} Feb 02 15:28:05 crc kubenswrapper[4733]: I0202 15:28:05.701100 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9gnj4" event={"ID":"501a8c1d-5254-4aac-b3c5-94a56b9e0fc0","Type":"ContainerStarted","Data":"8ba2ff8d2e36a00199ebce784dae046451eb38ee4648429dbbd7caa72327b898"} Feb 02 15:28:05 crc kubenswrapper[4733]: I0202 15:28:05.725611 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9gnj4" podStartSLOduration=2.304094115 podStartE2EDuration="5.725584325s" podCreationTimestamp="2026-02-02 15:28:00 +0000 UTC" firstStartedPulling="2026-02-02 15:28:01.669620126 +0000 UTC m=+765.121081484" lastFinishedPulling="2026-02-02 15:28:05.091110306 +0000 UTC m=+768.542571694" observedRunningTime="2026-02-02 15:28:05.717711474 +0000 UTC m=+769.169172902" watchObservedRunningTime="2026-02-02 15:28:05.725584325 +0000 UTC m=+769.177045723" Feb 02 15:28:11 crc kubenswrapper[4733]: I0202 15:28:11.165849 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9gnj4" Feb 02 15:28:11 crc kubenswrapper[4733]: I0202 15:28:11.166202 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9gnj4" Feb 02 15:28:12 crc kubenswrapper[4733]: I0202 15:28:12.224982 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9gnj4" podUID="501a8c1d-5254-4aac-b3c5-94a56b9e0fc0" containerName="registry-server" probeResult="failure" output=< Feb 02 15:28:12 crc kubenswrapper[4733]: timeout: failed to connect service ":50051" within 1s Feb 02 15:28:12 crc kubenswrapper[4733]: > Feb 02 15:28:14 crc kubenswrapper[4733]: I0202 15:28:14.340389 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-54b6695bd-mwgx2" Feb 02 15:28:14 crc kubenswrapper[4733]: I0202 15:28:14.990311 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-l6zg6"] Feb 02 15:28:14 crc kubenswrapper[4733]: I0202 15:28:14.993466 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-l6zg6" Feb 02 15:28:14 crc kubenswrapper[4733]: I0202 15:28:14.994757 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7df86c4f6c-5l6sj"] Feb 02 15:28:14 crc kubenswrapper[4733]: I0202 15:28:14.995797 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-5l6sj" Feb 02 15:28:14 crc kubenswrapper[4733]: I0202 15:28:14.996609 4733 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-jq4zx" Feb 02 15:28:14 crc kubenswrapper[4733]: I0202 15:28:14.996728 4733 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Feb 02 15:28:14 crc kubenswrapper[4733]: I0202 15:28:14.997735 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.000017 4733 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.007159 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7df86c4f6c-5l6sj"] Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.078751 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-7pnkj"] Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.080515 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-7pnkj" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.082642 4733 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-mh4nx" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.082953 4733 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.083329 4733 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.085447 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.100619 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-6968d8fdc4-h9fwf"] Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.101760 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6968d8fdc4-h9fwf" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.104120 4733 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.109278 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/70e35df4-ce17-4181-a64c-d7a70160d0e7-cert\") pod \"frr-k8s-webhook-server-7df86c4f6c-5l6sj\" (UID: \"70e35df4-ce17-4181-a64c-d7a70160d0e7\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-5l6sj" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.109564 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dw6w\" (UniqueName: \"kubernetes.io/projected/70e35df4-ce17-4181-a64c-d7a70160d0e7-kube-api-access-6dw6w\") pod \"frr-k8s-webhook-server-7df86c4f6c-5l6sj\" (UID: \"70e35df4-ce17-4181-a64c-d7a70160d0e7\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-5l6sj" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.109729 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/915914ae-77a5-4bf7-9d71-6207b62d2f50-frr-sockets\") pod \"frr-k8s-l6zg6\" (UID: \"915914ae-77a5-4bf7-9d71-6207b62d2f50\") " pod="metallb-system/frr-k8s-l6zg6" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.109782 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qs86m\" (UniqueName: \"kubernetes.io/projected/915914ae-77a5-4bf7-9d71-6207b62d2f50-kube-api-access-qs86m\") pod \"frr-k8s-l6zg6\" (UID: \"915914ae-77a5-4bf7-9d71-6207b62d2f50\") " pod="metallb-system/frr-k8s-l6zg6" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.109836 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/915914ae-77a5-4bf7-9d71-6207b62d2f50-metrics-certs\") pod \"frr-k8s-l6zg6\" (UID: \"915914ae-77a5-4bf7-9d71-6207b62d2f50\") " pod="metallb-system/frr-k8s-l6zg6" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.109896 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/915914ae-77a5-4bf7-9d71-6207b62d2f50-reloader\") pod \"frr-k8s-l6zg6\" (UID: \"915914ae-77a5-4bf7-9d71-6207b62d2f50\") " pod="metallb-system/frr-k8s-l6zg6" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.109920 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/915914ae-77a5-4bf7-9d71-6207b62d2f50-frr-conf\") pod \"frr-k8s-l6zg6\" (UID: \"915914ae-77a5-4bf7-9d71-6207b62d2f50\") " pod="metallb-system/frr-k8s-l6zg6" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.109970 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/915914ae-77a5-4bf7-9d71-6207b62d2f50-metrics\") pod \"frr-k8s-l6zg6\" (UID: \"915914ae-77a5-4bf7-9d71-6207b62d2f50\") " pod="metallb-system/frr-k8s-l6zg6" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.109992 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/915914ae-77a5-4bf7-9d71-6207b62d2f50-frr-startup\") pod \"frr-k8s-l6zg6\" (UID: \"915914ae-77a5-4bf7-9d71-6207b62d2f50\") " pod="metallb-system/frr-k8s-l6zg6" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.118473 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6968d8fdc4-h9fwf"] Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.211626 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/70e35df4-ce17-4181-a64c-d7a70160d0e7-cert\") pod \"frr-k8s-webhook-server-7df86c4f6c-5l6sj\" (UID: \"70e35df4-ce17-4181-a64c-d7a70160d0e7\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-5l6sj" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.211703 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpq7s\" (UniqueName: \"kubernetes.io/projected/222b6fa1-d59f-4779-8fdf-72c3fd3ea2cc-kube-api-access-cpq7s\") pod \"controller-6968d8fdc4-h9fwf\" (UID: \"222b6fa1-d59f-4779-8fdf-72c3fd3ea2cc\") " pod="metallb-system/controller-6968d8fdc4-h9fwf" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.211747 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dw6w\" (UniqueName: \"kubernetes.io/projected/70e35df4-ce17-4181-a64c-d7a70160d0e7-kube-api-access-6dw6w\") pod \"frr-k8s-webhook-server-7df86c4f6c-5l6sj\" (UID: \"70e35df4-ce17-4181-a64c-d7a70160d0e7\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-5l6sj" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.211812 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/915914ae-77a5-4bf7-9d71-6207b62d2f50-frr-sockets\") pod \"frr-k8s-l6zg6\" (UID: \"915914ae-77a5-4bf7-9d71-6207b62d2f50\") " pod="metallb-system/frr-k8s-l6zg6" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.211855 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qs86m\" (UniqueName: \"kubernetes.io/projected/915914ae-77a5-4bf7-9d71-6207b62d2f50-kube-api-access-qs86m\") pod \"frr-k8s-l6zg6\" (UID: \"915914ae-77a5-4bf7-9d71-6207b62d2f50\") " pod="metallb-system/frr-k8s-l6zg6" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.211896 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/915914ae-77a5-4bf7-9d71-6207b62d2f50-metrics-certs\") pod \"frr-k8s-l6zg6\" (UID: \"915914ae-77a5-4bf7-9d71-6207b62d2f50\") " pod="metallb-system/frr-k8s-l6zg6" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.211925 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/915914ae-77a5-4bf7-9d71-6207b62d2f50-reloader\") pod \"frr-k8s-l6zg6\" (UID: \"915914ae-77a5-4bf7-9d71-6207b62d2f50\") " pod="metallb-system/frr-k8s-l6zg6" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.211948 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1c5ad065-b35b-4784-89b6-72e8dd0a1df5-metrics-certs\") pod \"speaker-7pnkj\" (UID: \"1c5ad065-b35b-4784-89b6-72e8dd0a1df5\") " pod="metallb-system/speaker-7pnkj" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.211983 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/222b6fa1-d59f-4779-8fdf-72c3fd3ea2cc-metrics-certs\") pod \"controller-6968d8fdc4-h9fwf\" (UID: \"222b6fa1-d59f-4779-8fdf-72c3fd3ea2cc\") " pod="metallb-system/controller-6968d8fdc4-h9fwf" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.212005 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/915914ae-77a5-4bf7-9d71-6207b62d2f50-frr-conf\") pod \"frr-k8s-l6zg6\" (UID: \"915914ae-77a5-4bf7-9d71-6207b62d2f50\") " pod="metallb-system/frr-k8s-l6zg6" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.212031 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/1c5ad065-b35b-4784-89b6-72e8dd0a1df5-memberlist\") pod \"speaker-7pnkj\" (UID: \"1c5ad065-b35b-4784-89b6-72e8dd0a1df5\") " pod="metallb-system/speaker-7pnkj" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.212062 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/222b6fa1-d59f-4779-8fdf-72c3fd3ea2cc-cert\") pod \"controller-6968d8fdc4-h9fwf\" (UID: \"222b6fa1-d59f-4779-8fdf-72c3fd3ea2cc\") " pod="metallb-system/controller-6968d8fdc4-h9fwf" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.212087 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/915914ae-77a5-4bf7-9d71-6207b62d2f50-metrics\") pod \"frr-k8s-l6zg6\" (UID: \"915914ae-77a5-4bf7-9d71-6207b62d2f50\") " pod="metallb-system/frr-k8s-l6zg6" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.212109 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/1c5ad065-b35b-4784-89b6-72e8dd0a1df5-metallb-excludel2\") pod \"speaker-7pnkj\" (UID: \"1c5ad065-b35b-4784-89b6-72e8dd0a1df5\") " pod="metallb-system/speaker-7pnkj" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.212130 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/915914ae-77a5-4bf7-9d71-6207b62d2f50-frr-startup\") pod \"frr-k8s-l6zg6\" (UID: \"915914ae-77a5-4bf7-9d71-6207b62d2f50\") " pod="metallb-system/frr-k8s-l6zg6" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.212176 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9b5k\" (UniqueName: \"kubernetes.io/projected/1c5ad065-b35b-4784-89b6-72e8dd0a1df5-kube-api-access-b9b5k\") pod \"speaker-7pnkj\" (UID: \"1c5ad065-b35b-4784-89b6-72e8dd0a1df5\") " pod="metallb-system/speaker-7pnkj" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.212715 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/915914ae-77a5-4bf7-9d71-6207b62d2f50-frr-sockets\") pod \"frr-k8s-l6zg6\" (UID: \"915914ae-77a5-4bf7-9d71-6207b62d2f50\") " pod="metallb-system/frr-k8s-l6zg6" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.212922 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/915914ae-77a5-4bf7-9d71-6207b62d2f50-frr-conf\") pod \"frr-k8s-l6zg6\" (UID: \"915914ae-77a5-4bf7-9d71-6207b62d2f50\") " pod="metallb-system/frr-k8s-l6zg6" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.213054 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/915914ae-77a5-4bf7-9d71-6207b62d2f50-reloader\") pod \"frr-k8s-l6zg6\" (UID: \"915914ae-77a5-4bf7-9d71-6207b62d2f50\") " pod="metallb-system/frr-k8s-l6zg6" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.213068 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/915914ae-77a5-4bf7-9d71-6207b62d2f50-metrics\") pod \"frr-k8s-l6zg6\" (UID: \"915914ae-77a5-4bf7-9d71-6207b62d2f50\") " pod="metallb-system/frr-k8s-l6zg6" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.213428 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/915914ae-77a5-4bf7-9d71-6207b62d2f50-frr-startup\") pod \"frr-k8s-l6zg6\" (UID: \"915914ae-77a5-4bf7-9d71-6207b62d2f50\") " pod="metallb-system/frr-k8s-l6zg6" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.222127 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/70e35df4-ce17-4181-a64c-d7a70160d0e7-cert\") pod \"frr-k8s-webhook-server-7df86c4f6c-5l6sj\" (UID: \"70e35df4-ce17-4181-a64c-d7a70160d0e7\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-5l6sj" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.228151 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/915914ae-77a5-4bf7-9d71-6207b62d2f50-metrics-certs\") pod \"frr-k8s-l6zg6\" (UID: \"915914ae-77a5-4bf7-9d71-6207b62d2f50\") " pod="metallb-system/frr-k8s-l6zg6" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.241964 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dw6w\" (UniqueName: \"kubernetes.io/projected/70e35df4-ce17-4181-a64c-d7a70160d0e7-kube-api-access-6dw6w\") pod \"frr-k8s-webhook-server-7df86c4f6c-5l6sj\" (UID: \"70e35df4-ce17-4181-a64c-d7a70160d0e7\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-5l6sj" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.245844 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qs86m\" (UniqueName: \"kubernetes.io/projected/915914ae-77a5-4bf7-9d71-6207b62d2f50-kube-api-access-qs86m\") pod \"frr-k8s-l6zg6\" (UID: \"915914ae-77a5-4bf7-9d71-6207b62d2f50\") " pod="metallb-system/frr-k8s-l6zg6" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.313375 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/222b6fa1-d59f-4779-8fdf-72c3fd3ea2cc-cert\") pod \"controller-6968d8fdc4-h9fwf\" (UID: \"222b6fa1-d59f-4779-8fdf-72c3fd3ea2cc\") " pod="metallb-system/controller-6968d8fdc4-h9fwf" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.313430 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/1c5ad065-b35b-4784-89b6-72e8dd0a1df5-metallb-excludel2\") pod \"speaker-7pnkj\" (UID: \"1c5ad065-b35b-4784-89b6-72e8dd0a1df5\") " pod="metallb-system/speaker-7pnkj" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.313461 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9b5k\" (UniqueName: \"kubernetes.io/projected/1c5ad065-b35b-4784-89b6-72e8dd0a1df5-kube-api-access-b9b5k\") pod \"speaker-7pnkj\" (UID: \"1c5ad065-b35b-4784-89b6-72e8dd0a1df5\") " pod="metallb-system/speaker-7pnkj" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.313507 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpq7s\" (UniqueName: \"kubernetes.io/projected/222b6fa1-d59f-4779-8fdf-72c3fd3ea2cc-kube-api-access-cpq7s\") pod \"controller-6968d8fdc4-h9fwf\" (UID: \"222b6fa1-d59f-4779-8fdf-72c3fd3ea2cc\") " pod="metallb-system/controller-6968d8fdc4-h9fwf" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.313564 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1c5ad065-b35b-4784-89b6-72e8dd0a1df5-metrics-certs\") pod \"speaker-7pnkj\" (UID: \"1c5ad065-b35b-4784-89b6-72e8dd0a1df5\") " pod="metallb-system/speaker-7pnkj" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.313583 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/222b6fa1-d59f-4779-8fdf-72c3fd3ea2cc-metrics-certs\") pod \"controller-6968d8fdc4-h9fwf\" (UID: \"222b6fa1-d59f-4779-8fdf-72c3fd3ea2cc\") " pod="metallb-system/controller-6968d8fdc4-h9fwf" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.313609 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/1c5ad065-b35b-4784-89b6-72e8dd0a1df5-memberlist\") pod \"speaker-7pnkj\" (UID: \"1c5ad065-b35b-4784-89b6-72e8dd0a1df5\") " pod="metallb-system/speaker-7pnkj" Feb 02 15:28:15 crc kubenswrapper[4733]: E0202 15:28:15.313761 4733 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Feb 02 15:28:15 crc kubenswrapper[4733]: E0202 15:28:15.313823 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1c5ad065-b35b-4784-89b6-72e8dd0a1df5-memberlist podName:1c5ad065-b35b-4784-89b6-72e8dd0a1df5 nodeName:}" failed. No retries permitted until 2026-02-02 15:28:15.813800209 +0000 UTC m=+779.265261567 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/1c5ad065-b35b-4784-89b6-72e8dd0a1df5-memberlist") pod "speaker-7pnkj" (UID: "1c5ad065-b35b-4784-89b6-72e8dd0a1df5") : secret "metallb-memberlist" not found Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.314663 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/1c5ad065-b35b-4784-89b6-72e8dd0a1df5-metallb-excludel2\") pod \"speaker-7pnkj\" (UID: \"1c5ad065-b35b-4784-89b6-72e8dd0a1df5\") " pod="metallb-system/speaker-7pnkj" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.314819 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-l6zg6" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.320000 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1c5ad065-b35b-4784-89b6-72e8dd0a1df5-metrics-certs\") pod \"speaker-7pnkj\" (UID: \"1c5ad065-b35b-4784-89b6-72e8dd0a1df5\") " pod="metallb-system/speaker-7pnkj" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.322689 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/222b6fa1-d59f-4779-8fdf-72c3fd3ea2cc-cert\") pod \"controller-6968d8fdc4-h9fwf\" (UID: \"222b6fa1-d59f-4779-8fdf-72c3fd3ea2cc\") " pod="metallb-system/controller-6968d8fdc4-h9fwf" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.322865 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/222b6fa1-d59f-4779-8fdf-72c3fd3ea2cc-metrics-certs\") pod \"controller-6968d8fdc4-h9fwf\" (UID: \"222b6fa1-d59f-4779-8fdf-72c3fd3ea2cc\") " pod="metallb-system/controller-6968d8fdc4-h9fwf" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.323114 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-5l6sj" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.342133 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpq7s\" (UniqueName: \"kubernetes.io/projected/222b6fa1-d59f-4779-8fdf-72c3fd3ea2cc-kube-api-access-cpq7s\") pod \"controller-6968d8fdc4-h9fwf\" (UID: \"222b6fa1-d59f-4779-8fdf-72c3fd3ea2cc\") " pod="metallb-system/controller-6968d8fdc4-h9fwf" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.345669 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9b5k\" (UniqueName: \"kubernetes.io/projected/1c5ad065-b35b-4784-89b6-72e8dd0a1df5-kube-api-access-b9b5k\") pod \"speaker-7pnkj\" (UID: \"1c5ad065-b35b-4784-89b6-72e8dd0a1df5\") " pod="metallb-system/speaker-7pnkj" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.418266 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6968d8fdc4-h9fwf" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.511565 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7df86c4f6c-5l6sj"] Feb 02 15:28:15 crc kubenswrapper[4733]: W0202 15:28:15.524561 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod70e35df4_ce17_4181_a64c_d7a70160d0e7.slice/crio-16bd7a6387d165e69c9d3a8111267f418b190cae78b970a30569ad1a46fc84c6 WatchSource:0}: Error finding container 16bd7a6387d165e69c9d3a8111267f418b190cae78b970a30569ad1a46fc84c6: Status 404 returned error can't find the container with id 16bd7a6387d165e69c9d3a8111267f418b190cae78b970a30569ad1a46fc84c6 Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.600297 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6968d8fdc4-h9fwf"] Feb 02 15:28:15 crc kubenswrapper[4733]: W0202 15:28:15.603318 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod222b6fa1_d59f_4779_8fdf_72c3fd3ea2cc.slice/crio-70656699907b1d1c6f831ee1d5a19ec50b4d92103f8f9faf74e774377e5200c2 WatchSource:0}: Error finding container 70656699907b1d1c6f831ee1d5a19ec50b4d92103f8f9faf74e774377e5200c2: Status 404 returned error can't find the container with id 70656699907b1d1c6f831ee1d5a19ec50b4d92103f8f9faf74e774377e5200c2 Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.758849 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-l6zg6" event={"ID":"915914ae-77a5-4bf7-9d71-6207b62d2f50","Type":"ContainerStarted","Data":"950876bebaa1c11441f5e9e4dac71d0a9b41eea517a2d7645d8f09572602fc3f"} Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.759883 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-5l6sj" event={"ID":"70e35df4-ce17-4181-a64c-d7a70160d0e7","Type":"ContainerStarted","Data":"16bd7a6387d165e69c9d3a8111267f418b190cae78b970a30569ad1a46fc84c6"} Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.761438 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6968d8fdc4-h9fwf" event={"ID":"222b6fa1-d59f-4779-8fdf-72c3fd3ea2cc","Type":"ContainerStarted","Data":"65fbf352ffb1fe7a39babed0924121eabab675115748878099491e029fd02a14"} Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.761462 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6968d8fdc4-h9fwf" event={"ID":"222b6fa1-d59f-4779-8fdf-72c3fd3ea2cc","Type":"ContainerStarted","Data":"70656699907b1d1c6f831ee1d5a19ec50b4d92103f8f9faf74e774377e5200c2"} Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.829454 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/1c5ad065-b35b-4784-89b6-72e8dd0a1df5-memberlist\") pod \"speaker-7pnkj\" (UID: \"1c5ad065-b35b-4784-89b6-72e8dd0a1df5\") " pod="metallb-system/speaker-7pnkj" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.833730 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/1c5ad065-b35b-4784-89b6-72e8dd0a1df5-memberlist\") pod \"speaker-7pnkj\" (UID: \"1c5ad065-b35b-4784-89b6-72e8dd0a1df5\") " pod="metallb-system/speaker-7pnkj" Feb 02 15:28:15 crc kubenswrapper[4733]: I0202 15:28:15.993045 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-7pnkj" Feb 02 15:28:16 crc kubenswrapper[4733]: I0202 15:28:16.770310 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-7pnkj" event={"ID":"1c5ad065-b35b-4784-89b6-72e8dd0a1df5","Type":"ContainerStarted","Data":"42adf523f66a4f6dbbcc2de375cccc6a47590b822c51f1b625077ab3d9621404"} Feb 02 15:28:16 crc kubenswrapper[4733]: I0202 15:28:16.770666 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-7pnkj" event={"ID":"1c5ad065-b35b-4784-89b6-72e8dd0a1df5","Type":"ContainerStarted","Data":"167ff5bf7f1789880ebac7fab4a1b738c5e50b2c7ebc6483696fcb1512fa2d77"} Feb 02 15:28:17 crc kubenswrapper[4733]: I0202 15:28:17.650225 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-c5q6r"] Feb 02 15:28:17 crc kubenswrapper[4733]: I0202 15:28:17.651392 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c5q6r" Feb 02 15:28:17 crc kubenswrapper[4733]: I0202 15:28:17.659638 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-c5q6r"] Feb 02 15:28:17 crc kubenswrapper[4733]: I0202 15:28:17.666104 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qm4x\" (UniqueName: \"kubernetes.io/projected/3c3556a8-0e08-4f02-91cd-3b324e8dcbe8-kube-api-access-9qm4x\") pod \"community-operators-c5q6r\" (UID: \"3c3556a8-0e08-4f02-91cd-3b324e8dcbe8\") " pod="openshift-marketplace/community-operators-c5q6r" Feb 02 15:28:17 crc kubenswrapper[4733]: I0202 15:28:17.666407 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c3556a8-0e08-4f02-91cd-3b324e8dcbe8-utilities\") pod \"community-operators-c5q6r\" (UID: \"3c3556a8-0e08-4f02-91cd-3b324e8dcbe8\") " pod="openshift-marketplace/community-operators-c5q6r" Feb 02 15:28:17 crc kubenswrapper[4733]: I0202 15:28:17.666442 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c3556a8-0e08-4f02-91cd-3b324e8dcbe8-catalog-content\") pod \"community-operators-c5q6r\" (UID: \"3c3556a8-0e08-4f02-91cd-3b324e8dcbe8\") " pod="openshift-marketplace/community-operators-c5q6r" Feb 02 15:28:17 crc kubenswrapper[4733]: I0202 15:28:17.767829 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c3556a8-0e08-4f02-91cd-3b324e8dcbe8-utilities\") pod \"community-operators-c5q6r\" (UID: \"3c3556a8-0e08-4f02-91cd-3b324e8dcbe8\") " pod="openshift-marketplace/community-operators-c5q6r" Feb 02 15:28:17 crc kubenswrapper[4733]: I0202 15:28:17.767881 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c3556a8-0e08-4f02-91cd-3b324e8dcbe8-catalog-content\") pod \"community-operators-c5q6r\" (UID: \"3c3556a8-0e08-4f02-91cd-3b324e8dcbe8\") " pod="openshift-marketplace/community-operators-c5q6r" Feb 02 15:28:17 crc kubenswrapper[4733]: I0202 15:28:17.767912 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qm4x\" (UniqueName: \"kubernetes.io/projected/3c3556a8-0e08-4f02-91cd-3b324e8dcbe8-kube-api-access-9qm4x\") pod \"community-operators-c5q6r\" (UID: \"3c3556a8-0e08-4f02-91cd-3b324e8dcbe8\") " pod="openshift-marketplace/community-operators-c5q6r" Feb 02 15:28:17 crc kubenswrapper[4733]: I0202 15:28:17.768409 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c3556a8-0e08-4f02-91cd-3b324e8dcbe8-catalog-content\") pod \"community-operators-c5q6r\" (UID: \"3c3556a8-0e08-4f02-91cd-3b324e8dcbe8\") " pod="openshift-marketplace/community-operators-c5q6r" Feb 02 15:28:17 crc kubenswrapper[4733]: I0202 15:28:17.768508 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c3556a8-0e08-4f02-91cd-3b324e8dcbe8-utilities\") pod \"community-operators-c5q6r\" (UID: \"3c3556a8-0e08-4f02-91cd-3b324e8dcbe8\") " pod="openshift-marketplace/community-operators-c5q6r" Feb 02 15:28:17 crc kubenswrapper[4733]: I0202 15:28:17.802190 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qm4x\" (UniqueName: \"kubernetes.io/projected/3c3556a8-0e08-4f02-91cd-3b324e8dcbe8-kube-api-access-9qm4x\") pod \"community-operators-c5q6r\" (UID: \"3c3556a8-0e08-4f02-91cd-3b324e8dcbe8\") " pod="openshift-marketplace/community-operators-c5q6r" Feb 02 15:28:17 crc kubenswrapper[4733]: I0202 15:28:17.994291 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c5q6r" Feb 02 15:28:19 crc kubenswrapper[4733]: I0202 15:28:19.463750 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-c5q6r"] Feb 02 15:28:19 crc kubenswrapper[4733]: I0202 15:28:19.795661 4733 generic.go:334] "Generic (PLEG): container finished" podID="3c3556a8-0e08-4f02-91cd-3b324e8dcbe8" containerID="2cf1e253ed16db53041b1c8290ee4c837ef70cedade2f6511a598e9efa417409" exitCode=0 Feb 02 15:28:19 crc kubenswrapper[4733]: I0202 15:28:19.795729 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c5q6r" event={"ID":"3c3556a8-0e08-4f02-91cd-3b324e8dcbe8","Type":"ContainerDied","Data":"2cf1e253ed16db53041b1c8290ee4c837ef70cedade2f6511a598e9efa417409"} Feb 02 15:28:19 crc kubenswrapper[4733]: I0202 15:28:19.796000 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c5q6r" event={"ID":"3c3556a8-0e08-4f02-91cd-3b324e8dcbe8","Type":"ContainerStarted","Data":"5f4c0f856fec2e9b21315fde66b40fa5baaf18d73faa1c6bfa433a303f2723e8"} Feb 02 15:28:20 crc kubenswrapper[4733]: I0202 15:28:20.805520 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-7pnkj" event={"ID":"1c5ad065-b35b-4784-89b6-72e8dd0a1df5","Type":"ContainerStarted","Data":"c64dc2b8de5cea890f1500f1ea9862cc042565050c09b869d5826e62fc2e50d3"} Feb 02 15:28:20 crc kubenswrapper[4733]: I0202 15:28:20.805875 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-7pnkj" Feb 02 15:28:20 crc kubenswrapper[4733]: I0202 15:28:20.807822 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6968d8fdc4-h9fwf" event={"ID":"222b6fa1-d59f-4779-8fdf-72c3fd3ea2cc","Type":"ContainerStarted","Data":"afa318be02b5cbd3cef05e9a2d7b72aad0e98be21987777e5e94f2625a147042"} Feb 02 15:28:20 crc kubenswrapper[4733]: I0202 15:28:20.808007 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-6968d8fdc4-h9fwf" Feb 02 15:28:20 crc kubenswrapper[4733]: I0202 15:28:20.827441 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-7pnkj" podStartSLOduration=2.263910649 podStartE2EDuration="5.827419249s" podCreationTimestamp="2026-02-02 15:28:15 +0000 UTC" firstStartedPulling="2026-02-02 15:28:16.320142091 +0000 UTC m=+779.771603459" lastFinishedPulling="2026-02-02 15:28:19.883650701 +0000 UTC m=+783.335112059" observedRunningTime="2026-02-02 15:28:20.826520916 +0000 UTC m=+784.277982294" watchObservedRunningTime="2026-02-02 15:28:20.827419249 +0000 UTC m=+784.278880607" Feb 02 15:28:20 crc kubenswrapper[4733]: I0202 15:28:20.849523 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-6968d8fdc4-h9fwf" podStartSLOduration=1.690308278 podStartE2EDuration="5.849505846s" podCreationTimestamp="2026-02-02 15:28:15 +0000 UTC" firstStartedPulling="2026-02-02 15:28:15.705039066 +0000 UTC m=+779.156500424" lastFinishedPulling="2026-02-02 15:28:19.864236634 +0000 UTC m=+783.315697992" observedRunningTime="2026-02-02 15:28:20.844061774 +0000 UTC m=+784.295523152" watchObservedRunningTime="2026-02-02 15:28:20.849505846 +0000 UTC m=+784.300967214" Feb 02 15:28:21 crc kubenswrapper[4733]: I0202 15:28:21.215715 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9gnj4" Feb 02 15:28:21 crc kubenswrapper[4733]: I0202 15:28:21.296716 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9gnj4" Feb 02 15:28:22 crc kubenswrapper[4733]: I0202 15:28:22.027471 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9gnj4"] Feb 02 15:28:22 crc kubenswrapper[4733]: I0202 15:28:22.824807 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9gnj4" podUID="501a8c1d-5254-4aac-b3c5-94a56b9e0fc0" containerName="registry-server" containerID="cri-o://8ba2ff8d2e36a00199ebce784dae046451eb38ee4648429dbbd7caa72327b898" gracePeriod=2 Feb 02 15:28:23 crc kubenswrapper[4733]: I0202 15:28:23.208951 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9gnj4" Feb 02 15:28:23 crc kubenswrapper[4733]: I0202 15:28:23.256361 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/501a8c1d-5254-4aac-b3c5-94a56b9e0fc0-utilities\") pod \"501a8c1d-5254-4aac-b3c5-94a56b9e0fc0\" (UID: \"501a8c1d-5254-4aac-b3c5-94a56b9e0fc0\") " Feb 02 15:28:23 crc kubenswrapper[4733]: I0202 15:28:23.256474 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-58vdc\" (UniqueName: \"kubernetes.io/projected/501a8c1d-5254-4aac-b3c5-94a56b9e0fc0-kube-api-access-58vdc\") pod \"501a8c1d-5254-4aac-b3c5-94a56b9e0fc0\" (UID: \"501a8c1d-5254-4aac-b3c5-94a56b9e0fc0\") " Feb 02 15:28:23 crc kubenswrapper[4733]: I0202 15:28:23.256510 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/501a8c1d-5254-4aac-b3c5-94a56b9e0fc0-catalog-content\") pod \"501a8c1d-5254-4aac-b3c5-94a56b9e0fc0\" (UID: \"501a8c1d-5254-4aac-b3c5-94a56b9e0fc0\") " Feb 02 15:28:23 crc kubenswrapper[4733]: I0202 15:28:23.269909 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/501a8c1d-5254-4aac-b3c5-94a56b9e0fc0-kube-api-access-58vdc" (OuterVolumeSpecName: "kube-api-access-58vdc") pod "501a8c1d-5254-4aac-b3c5-94a56b9e0fc0" (UID: "501a8c1d-5254-4aac-b3c5-94a56b9e0fc0"). InnerVolumeSpecName "kube-api-access-58vdc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:28:23 crc kubenswrapper[4733]: I0202 15:28:23.273657 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/501a8c1d-5254-4aac-b3c5-94a56b9e0fc0-utilities" (OuterVolumeSpecName: "utilities") pod "501a8c1d-5254-4aac-b3c5-94a56b9e0fc0" (UID: "501a8c1d-5254-4aac-b3c5-94a56b9e0fc0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:28:23 crc kubenswrapper[4733]: I0202 15:28:23.357921 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-58vdc\" (UniqueName: \"kubernetes.io/projected/501a8c1d-5254-4aac-b3c5-94a56b9e0fc0-kube-api-access-58vdc\") on node \"crc\" DevicePath \"\"" Feb 02 15:28:23 crc kubenswrapper[4733]: I0202 15:28:23.357971 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/501a8c1d-5254-4aac-b3c5-94a56b9e0fc0-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 15:28:23 crc kubenswrapper[4733]: I0202 15:28:23.406142 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/501a8c1d-5254-4aac-b3c5-94a56b9e0fc0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "501a8c1d-5254-4aac-b3c5-94a56b9e0fc0" (UID: "501a8c1d-5254-4aac-b3c5-94a56b9e0fc0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:28:23 crc kubenswrapper[4733]: I0202 15:28:23.458779 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/501a8c1d-5254-4aac-b3c5-94a56b9e0fc0-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 15:28:23 crc kubenswrapper[4733]: I0202 15:28:23.830536 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-5l6sj" event={"ID":"70e35df4-ce17-4181-a64c-d7a70160d0e7","Type":"ContainerStarted","Data":"b7220c2a5620106275d3b03bdd5bf131dfbbcf95698387555ca1381b52196ab3"} Feb 02 15:28:23 crc kubenswrapper[4733]: I0202 15:28:23.830668 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-5l6sj" Feb 02 15:28:23 crc kubenswrapper[4733]: I0202 15:28:23.831970 4733 generic.go:334] "Generic (PLEG): container finished" podID="501a8c1d-5254-4aac-b3c5-94a56b9e0fc0" containerID="8ba2ff8d2e36a00199ebce784dae046451eb38ee4648429dbbd7caa72327b898" exitCode=0 Feb 02 15:28:23 crc kubenswrapper[4733]: I0202 15:28:23.832024 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9gnj4" event={"ID":"501a8c1d-5254-4aac-b3c5-94a56b9e0fc0","Type":"ContainerDied","Data":"8ba2ff8d2e36a00199ebce784dae046451eb38ee4648429dbbd7caa72327b898"} Feb 02 15:28:23 crc kubenswrapper[4733]: I0202 15:28:23.832038 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9gnj4" Feb 02 15:28:23 crc kubenswrapper[4733]: I0202 15:28:23.832048 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9gnj4" event={"ID":"501a8c1d-5254-4aac-b3c5-94a56b9e0fc0","Type":"ContainerDied","Data":"ea087ab9d4010689b38db8c58c6f1fd8316053ff6d447d24f390ff9d36fefd07"} Feb 02 15:28:23 crc kubenswrapper[4733]: I0202 15:28:23.832079 4733 scope.go:117] "RemoveContainer" containerID="8ba2ff8d2e36a00199ebce784dae046451eb38ee4648429dbbd7caa72327b898" Feb 02 15:28:23 crc kubenswrapper[4733]: I0202 15:28:23.834774 4733 generic.go:334] "Generic (PLEG): container finished" podID="3c3556a8-0e08-4f02-91cd-3b324e8dcbe8" containerID="cf999c48149fa874d4f30fdb43f55656b7477f1478f5ff0263514c147101f7bb" exitCode=0 Feb 02 15:28:23 crc kubenswrapper[4733]: I0202 15:28:23.834822 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c5q6r" event={"ID":"3c3556a8-0e08-4f02-91cd-3b324e8dcbe8","Type":"ContainerDied","Data":"cf999c48149fa874d4f30fdb43f55656b7477f1478f5ff0263514c147101f7bb"} Feb 02 15:28:23 crc kubenswrapper[4733]: I0202 15:28:23.839144 4733 generic.go:334] "Generic (PLEG): container finished" podID="915914ae-77a5-4bf7-9d71-6207b62d2f50" containerID="3afe9717c891f08a608dbd0e94ac28e91a11054f8ab5cd8a08a38afb2b6a8773" exitCode=0 Feb 02 15:28:23 crc kubenswrapper[4733]: I0202 15:28:23.839194 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-l6zg6" event={"ID":"915914ae-77a5-4bf7-9d71-6207b62d2f50","Type":"ContainerDied","Data":"3afe9717c891f08a608dbd0e94ac28e91a11054f8ab5cd8a08a38afb2b6a8773"} Feb 02 15:28:23 crc kubenswrapper[4733]: I0202 15:28:23.845947 4733 scope.go:117] "RemoveContainer" containerID="3c65c98c0c190da476bcee6ca1e2c1bc0b5f9be0c1dc50cc9bcf405e5dd47c68" Feb 02 15:28:23 crc kubenswrapper[4733]: I0202 15:28:23.865816 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-5l6sj" podStartSLOduration=2.385841236 podStartE2EDuration="9.865797661s" podCreationTimestamp="2026-02-02 15:28:14 +0000 UTC" firstStartedPulling="2026-02-02 15:28:15.526692713 +0000 UTC m=+778.978154071" lastFinishedPulling="2026-02-02 15:28:23.006649148 +0000 UTC m=+786.458110496" observedRunningTime="2026-02-02 15:28:23.860612725 +0000 UTC m=+787.312074093" watchObservedRunningTime="2026-02-02 15:28:23.865797661 +0000 UTC m=+787.317259019" Feb 02 15:28:23 crc kubenswrapper[4733]: I0202 15:28:23.867269 4733 scope.go:117] "RemoveContainer" containerID="c8901e6c33323ad3b5bdbe10291cb7d14bed6ddaee9abd04ef234f19fa80684d" Feb 02 15:28:23 crc kubenswrapper[4733]: I0202 15:28:23.877845 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9gnj4"] Feb 02 15:28:23 crc kubenswrapper[4733]: I0202 15:28:23.882539 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9gnj4"] Feb 02 15:28:23 crc kubenswrapper[4733]: I0202 15:28:23.888118 4733 scope.go:117] "RemoveContainer" containerID="8ba2ff8d2e36a00199ebce784dae046451eb38ee4648429dbbd7caa72327b898" Feb 02 15:28:23 crc kubenswrapper[4733]: E0202 15:28:23.889703 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ba2ff8d2e36a00199ebce784dae046451eb38ee4648429dbbd7caa72327b898\": container with ID starting with 8ba2ff8d2e36a00199ebce784dae046451eb38ee4648429dbbd7caa72327b898 not found: ID does not exist" containerID="8ba2ff8d2e36a00199ebce784dae046451eb38ee4648429dbbd7caa72327b898" Feb 02 15:28:23 crc kubenswrapper[4733]: I0202 15:28:23.889856 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ba2ff8d2e36a00199ebce784dae046451eb38ee4648429dbbd7caa72327b898"} err="failed to get container status \"8ba2ff8d2e36a00199ebce784dae046451eb38ee4648429dbbd7caa72327b898\": rpc error: code = NotFound desc = could not find container \"8ba2ff8d2e36a00199ebce784dae046451eb38ee4648429dbbd7caa72327b898\": container with ID starting with 8ba2ff8d2e36a00199ebce784dae046451eb38ee4648429dbbd7caa72327b898 not found: ID does not exist" Feb 02 15:28:23 crc kubenswrapper[4733]: I0202 15:28:23.889975 4733 scope.go:117] "RemoveContainer" containerID="3c65c98c0c190da476bcee6ca1e2c1bc0b5f9be0c1dc50cc9bcf405e5dd47c68" Feb 02 15:28:23 crc kubenswrapper[4733]: E0202 15:28:23.890663 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c65c98c0c190da476bcee6ca1e2c1bc0b5f9be0c1dc50cc9bcf405e5dd47c68\": container with ID starting with 3c65c98c0c190da476bcee6ca1e2c1bc0b5f9be0c1dc50cc9bcf405e5dd47c68 not found: ID does not exist" containerID="3c65c98c0c190da476bcee6ca1e2c1bc0b5f9be0c1dc50cc9bcf405e5dd47c68" Feb 02 15:28:23 crc kubenswrapper[4733]: I0202 15:28:23.890711 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c65c98c0c190da476bcee6ca1e2c1bc0b5f9be0c1dc50cc9bcf405e5dd47c68"} err="failed to get container status \"3c65c98c0c190da476bcee6ca1e2c1bc0b5f9be0c1dc50cc9bcf405e5dd47c68\": rpc error: code = NotFound desc = could not find container \"3c65c98c0c190da476bcee6ca1e2c1bc0b5f9be0c1dc50cc9bcf405e5dd47c68\": container with ID starting with 3c65c98c0c190da476bcee6ca1e2c1bc0b5f9be0c1dc50cc9bcf405e5dd47c68 not found: ID does not exist" Feb 02 15:28:23 crc kubenswrapper[4733]: I0202 15:28:23.890746 4733 scope.go:117] "RemoveContainer" containerID="c8901e6c33323ad3b5bdbe10291cb7d14bed6ddaee9abd04ef234f19fa80684d" Feb 02 15:28:23 crc kubenswrapper[4733]: E0202 15:28:23.891142 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8901e6c33323ad3b5bdbe10291cb7d14bed6ddaee9abd04ef234f19fa80684d\": container with ID starting with c8901e6c33323ad3b5bdbe10291cb7d14bed6ddaee9abd04ef234f19fa80684d not found: ID does not exist" containerID="c8901e6c33323ad3b5bdbe10291cb7d14bed6ddaee9abd04ef234f19fa80684d" Feb 02 15:28:23 crc kubenswrapper[4733]: I0202 15:28:23.891460 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8901e6c33323ad3b5bdbe10291cb7d14bed6ddaee9abd04ef234f19fa80684d"} err="failed to get container status \"c8901e6c33323ad3b5bdbe10291cb7d14bed6ddaee9abd04ef234f19fa80684d\": rpc error: code = NotFound desc = could not find container \"c8901e6c33323ad3b5bdbe10291cb7d14bed6ddaee9abd04ef234f19fa80684d\": container with ID starting with c8901e6c33323ad3b5bdbe10291cb7d14bed6ddaee9abd04ef234f19fa80684d not found: ID does not exist" Feb 02 15:28:24 crc kubenswrapper[4733]: I0202 15:28:24.848989 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c5q6r" event={"ID":"3c3556a8-0e08-4f02-91cd-3b324e8dcbe8","Type":"ContainerStarted","Data":"d9dbc639cf53a165aa8dfdf5f9be047ba16af94c682cb64eae180ac3dbee03a8"} Feb 02 15:28:24 crc kubenswrapper[4733]: I0202 15:28:24.854313 4733 generic.go:334] "Generic (PLEG): container finished" podID="915914ae-77a5-4bf7-9d71-6207b62d2f50" containerID="f1b0b7ea0bd6269b0b0a722456b658df8fe1794f4df539ed6ec6925479d4a2a7" exitCode=0 Feb 02 15:28:24 crc kubenswrapper[4733]: I0202 15:28:24.854427 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-l6zg6" event={"ID":"915914ae-77a5-4bf7-9d71-6207b62d2f50","Type":"ContainerDied","Data":"f1b0b7ea0bd6269b0b0a722456b658df8fe1794f4df539ed6ec6925479d4a2a7"} Feb 02 15:28:24 crc kubenswrapper[4733]: I0202 15:28:24.881620 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-c5q6r" podStartSLOduration=3.453055874 podStartE2EDuration="7.881592908s" podCreationTimestamp="2026-02-02 15:28:17 +0000 UTC" firstStartedPulling="2026-02-02 15:28:19.836993923 +0000 UTC m=+783.288455301" lastFinishedPulling="2026-02-02 15:28:24.265530977 +0000 UTC m=+787.716992335" observedRunningTime="2026-02-02 15:28:24.874901554 +0000 UTC m=+788.326362932" watchObservedRunningTime="2026-02-02 15:28:24.881592908 +0000 UTC m=+788.333054296" Feb 02 15:28:25 crc kubenswrapper[4733]: I0202 15:28:25.265305 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="501a8c1d-5254-4aac-b3c5-94a56b9e0fc0" path="/var/lib/kubelet/pods/501a8c1d-5254-4aac-b3c5-94a56b9e0fc0/volumes" Feb 02 15:28:25 crc kubenswrapper[4733]: I0202 15:28:25.425725 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-6968d8fdc4-h9fwf" Feb 02 15:28:25 crc kubenswrapper[4733]: I0202 15:28:25.862297 4733 generic.go:334] "Generic (PLEG): container finished" podID="915914ae-77a5-4bf7-9d71-6207b62d2f50" containerID="5175f6a1a7951e4eac477e09a6a9b6d45f2427366e7d7da09e0e8e070b716205" exitCode=0 Feb 02 15:28:25 crc kubenswrapper[4733]: I0202 15:28:25.862351 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-l6zg6" event={"ID":"915914ae-77a5-4bf7-9d71-6207b62d2f50","Type":"ContainerDied","Data":"5175f6a1a7951e4eac477e09a6a9b6d45f2427366e7d7da09e0e8e070b716205"} Feb 02 15:28:26 crc kubenswrapper[4733]: I0202 15:28:26.872768 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-l6zg6" event={"ID":"915914ae-77a5-4bf7-9d71-6207b62d2f50","Type":"ContainerStarted","Data":"ba002ea63792e1e28057f46f1e0658188fa1ec68b6a208b54f14cf0dafa80bef"} Feb 02 15:28:26 crc kubenswrapper[4733]: I0202 15:28:26.872827 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-l6zg6" event={"ID":"915914ae-77a5-4bf7-9d71-6207b62d2f50","Type":"ContainerStarted","Data":"8a1fd797a6a717e3ea8bf7ff0c4ebb7bf85af94f47c82cdbef3054ee9348cd87"} Feb 02 15:28:26 crc kubenswrapper[4733]: I0202 15:28:26.872841 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-l6zg6" event={"ID":"915914ae-77a5-4bf7-9d71-6207b62d2f50","Type":"ContainerStarted","Data":"a6fa3d82d947a072b3fff20692693d0f4a586ef0df52823f80eaeefbdf2d8bc7"} Feb 02 15:28:26 crc kubenswrapper[4733]: I0202 15:28:26.872854 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-l6zg6" event={"ID":"915914ae-77a5-4bf7-9d71-6207b62d2f50","Type":"ContainerStarted","Data":"78b8dc6ada9bc6dd05c53cd41421fc7070cb0a8f07b00ff300262ed732c08a0e"} Feb 02 15:28:26 crc kubenswrapper[4733]: I0202 15:28:26.872864 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-l6zg6" event={"ID":"915914ae-77a5-4bf7-9d71-6207b62d2f50","Type":"ContainerStarted","Data":"8b16a50f218999f59fe21e3dd9d4cc0ecc607f35e159a10c71feda3984a2a55f"} Feb 02 15:28:26 crc kubenswrapper[4733]: I0202 15:28:26.872873 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-l6zg6" event={"ID":"915914ae-77a5-4bf7-9d71-6207b62d2f50","Type":"ContainerStarted","Data":"03724d1060eeb7a7bbf1f3394639fa74ec7c70ba68818817f1829d4dbc6ad064"} Feb 02 15:28:26 crc kubenswrapper[4733]: I0202 15:28:26.872896 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-l6zg6" Feb 02 15:28:27 crc kubenswrapper[4733]: I0202 15:28:27.995178 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-c5q6r" Feb 02 15:28:27 crc kubenswrapper[4733]: I0202 15:28:27.995519 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-c5q6r" Feb 02 15:28:28 crc kubenswrapper[4733]: I0202 15:28:28.050097 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-c5q6r" Feb 02 15:28:28 crc kubenswrapper[4733]: I0202 15:28:28.081337 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-l6zg6" podStartSLOduration=6.5848356710000004 podStartE2EDuration="14.081311697s" podCreationTimestamp="2026-02-02 15:28:14 +0000 UTC" firstStartedPulling="2026-02-02 15:28:15.481376751 +0000 UTC m=+778.932838109" lastFinishedPulling="2026-02-02 15:28:22.977852747 +0000 UTC m=+786.429314135" observedRunningTime="2026-02-02 15:28:26.89877778 +0000 UTC m=+790.350239148" watchObservedRunningTime="2026-02-02 15:28:28.081311697 +0000 UTC m=+791.532773055" Feb 02 15:28:28 crc kubenswrapper[4733]: I0202 15:28:28.240219 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-kn8hn"] Feb 02 15:28:28 crc kubenswrapper[4733]: E0202 15:28:28.240541 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="501a8c1d-5254-4aac-b3c5-94a56b9e0fc0" containerName="registry-server" Feb 02 15:28:28 crc kubenswrapper[4733]: I0202 15:28:28.240559 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="501a8c1d-5254-4aac-b3c5-94a56b9e0fc0" containerName="registry-server" Feb 02 15:28:28 crc kubenswrapper[4733]: E0202 15:28:28.240572 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="501a8c1d-5254-4aac-b3c5-94a56b9e0fc0" containerName="extract-content" Feb 02 15:28:28 crc kubenswrapper[4733]: I0202 15:28:28.240581 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="501a8c1d-5254-4aac-b3c5-94a56b9e0fc0" containerName="extract-content" Feb 02 15:28:28 crc kubenswrapper[4733]: E0202 15:28:28.240602 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="501a8c1d-5254-4aac-b3c5-94a56b9e0fc0" containerName="extract-utilities" Feb 02 15:28:28 crc kubenswrapper[4733]: I0202 15:28:28.240611 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="501a8c1d-5254-4aac-b3c5-94a56b9e0fc0" containerName="extract-utilities" Feb 02 15:28:28 crc kubenswrapper[4733]: I0202 15:28:28.240755 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="501a8c1d-5254-4aac-b3c5-94a56b9e0fc0" containerName="registry-server" Feb 02 15:28:28 crc kubenswrapper[4733]: I0202 15:28:28.241723 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kn8hn" Feb 02 15:28:28 crc kubenswrapper[4733]: I0202 15:28:28.257077 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kn8hn"] Feb 02 15:28:28 crc kubenswrapper[4733]: I0202 15:28:28.337919 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afb40853-b770-44fb-a848-d23ea55ef184-utilities\") pod \"redhat-marketplace-kn8hn\" (UID: \"afb40853-b770-44fb-a848-d23ea55ef184\") " pod="openshift-marketplace/redhat-marketplace-kn8hn" Feb 02 15:28:28 crc kubenswrapper[4733]: I0202 15:28:28.338267 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chvlq\" (UniqueName: \"kubernetes.io/projected/afb40853-b770-44fb-a848-d23ea55ef184-kube-api-access-chvlq\") pod \"redhat-marketplace-kn8hn\" (UID: \"afb40853-b770-44fb-a848-d23ea55ef184\") " pod="openshift-marketplace/redhat-marketplace-kn8hn" Feb 02 15:28:28 crc kubenswrapper[4733]: I0202 15:28:28.338387 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afb40853-b770-44fb-a848-d23ea55ef184-catalog-content\") pod \"redhat-marketplace-kn8hn\" (UID: \"afb40853-b770-44fb-a848-d23ea55ef184\") " pod="openshift-marketplace/redhat-marketplace-kn8hn" Feb 02 15:28:28 crc kubenswrapper[4733]: I0202 15:28:28.439427 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chvlq\" (UniqueName: \"kubernetes.io/projected/afb40853-b770-44fb-a848-d23ea55ef184-kube-api-access-chvlq\") pod \"redhat-marketplace-kn8hn\" (UID: \"afb40853-b770-44fb-a848-d23ea55ef184\") " pod="openshift-marketplace/redhat-marketplace-kn8hn" Feb 02 15:28:28 crc kubenswrapper[4733]: I0202 15:28:28.439493 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afb40853-b770-44fb-a848-d23ea55ef184-catalog-content\") pod \"redhat-marketplace-kn8hn\" (UID: \"afb40853-b770-44fb-a848-d23ea55ef184\") " pod="openshift-marketplace/redhat-marketplace-kn8hn" Feb 02 15:28:28 crc kubenswrapper[4733]: I0202 15:28:28.439602 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afb40853-b770-44fb-a848-d23ea55ef184-utilities\") pod \"redhat-marketplace-kn8hn\" (UID: \"afb40853-b770-44fb-a848-d23ea55ef184\") " pod="openshift-marketplace/redhat-marketplace-kn8hn" Feb 02 15:28:28 crc kubenswrapper[4733]: I0202 15:28:28.440211 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afb40853-b770-44fb-a848-d23ea55ef184-utilities\") pod \"redhat-marketplace-kn8hn\" (UID: \"afb40853-b770-44fb-a848-d23ea55ef184\") " pod="openshift-marketplace/redhat-marketplace-kn8hn" Feb 02 15:28:28 crc kubenswrapper[4733]: I0202 15:28:28.440413 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afb40853-b770-44fb-a848-d23ea55ef184-catalog-content\") pod \"redhat-marketplace-kn8hn\" (UID: \"afb40853-b770-44fb-a848-d23ea55ef184\") " pod="openshift-marketplace/redhat-marketplace-kn8hn" Feb 02 15:28:28 crc kubenswrapper[4733]: I0202 15:28:28.466204 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chvlq\" (UniqueName: \"kubernetes.io/projected/afb40853-b770-44fb-a848-d23ea55ef184-kube-api-access-chvlq\") pod \"redhat-marketplace-kn8hn\" (UID: \"afb40853-b770-44fb-a848-d23ea55ef184\") " pod="openshift-marketplace/redhat-marketplace-kn8hn" Feb 02 15:28:28 crc kubenswrapper[4733]: I0202 15:28:28.558936 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kn8hn" Feb 02 15:28:28 crc kubenswrapper[4733]: I0202 15:28:28.764495 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kn8hn"] Feb 02 15:28:28 crc kubenswrapper[4733]: W0202 15:28:28.793349 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podafb40853_b770_44fb_a848_d23ea55ef184.slice/crio-bc96c332030f5f9655fbd440f2b8414bfd2c42cf55f6ebc5d1934f427ef418c2 WatchSource:0}: Error finding container bc96c332030f5f9655fbd440f2b8414bfd2c42cf55f6ebc5d1934f427ef418c2: Status 404 returned error can't find the container with id bc96c332030f5f9655fbd440f2b8414bfd2c42cf55f6ebc5d1934f427ef418c2 Feb 02 15:28:28 crc kubenswrapper[4733]: I0202 15:28:28.888341 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kn8hn" event={"ID":"afb40853-b770-44fb-a848-d23ea55ef184","Type":"ContainerStarted","Data":"bc96c332030f5f9655fbd440f2b8414bfd2c42cf55f6ebc5d1934f427ef418c2"} Feb 02 15:28:29 crc kubenswrapper[4733]: I0202 15:28:29.897080 4733 generic.go:334] "Generic (PLEG): container finished" podID="afb40853-b770-44fb-a848-d23ea55ef184" containerID="7aae18ef19919774f2cde8911e5a4460acd630dac3077ffb2d26ad1ede6189d3" exitCode=0 Feb 02 15:28:29 crc kubenswrapper[4733]: I0202 15:28:29.897250 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kn8hn" event={"ID":"afb40853-b770-44fb-a848-d23ea55ef184","Type":"ContainerDied","Data":"7aae18ef19919774f2cde8911e5a4460acd630dac3077ffb2d26ad1ede6189d3"} Feb 02 15:28:30 crc kubenswrapper[4733]: I0202 15:28:30.315828 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-l6zg6" Feb 02 15:28:30 crc kubenswrapper[4733]: I0202 15:28:30.351475 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-l6zg6" Feb 02 15:28:31 crc kubenswrapper[4733]: I0202 15:28:31.915052 4733 generic.go:334] "Generic (PLEG): container finished" podID="afb40853-b770-44fb-a848-d23ea55ef184" containerID="39f1449c1004537b0605cd0c1c497cd15c0936cef5c4bc3fd603b8abb6f9a684" exitCode=0 Feb 02 15:28:31 crc kubenswrapper[4733]: I0202 15:28:31.915220 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kn8hn" event={"ID":"afb40853-b770-44fb-a848-d23ea55ef184","Type":"ContainerDied","Data":"39f1449c1004537b0605cd0c1c497cd15c0936cef5c4bc3fd603b8abb6f9a684"} Feb 02 15:28:32 crc kubenswrapper[4733]: I0202 15:28:32.951544 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kn8hn" event={"ID":"afb40853-b770-44fb-a848-d23ea55ef184","Type":"ContainerStarted","Data":"b638997e132de496c2331dc3e264df17c6ba3e287577ba916d0dfd55c30417e5"} Feb 02 15:28:32 crc kubenswrapper[4733]: I0202 15:28:32.973353 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-kn8hn" podStartSLOduration=2.3013313 podStartE2EDuration="4.973333283s" podCreationTimestamp="2026-02-02 15:28:28 +0000 UTC" firstStartedPulling="2026-02-02 15:28:29.899439226 +0000 UTC m=+793.350900594" lastFinishedPulling="2026-02-02 15:28:32.571441179 +0000 UTC m=+796.022902577" observedRunningTime="2026-02-02 15:28:32.970591292 +0000 UTC m=+796.422052650" watchObservedRunningTime="2026-02-02 15:28:32.973333283 +0000 UTC m=+796.424794641" Feb 02 15:28:34 crc kubenswrapper[4733]: I0202 15:28:34.987909 4733 patch_prober.go:28] interesting pod/machine-config-daemon-zgsv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 15:28:34 crc kubenswrapper[4733]: I0202 15:28:34.987997 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 15:28:35 crc kubenswrapper[4733]: I0202 15:28:35.319689 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-l6zg6" Feb 02 15:28:35 crc kubenswrapper[4733]: I0202 15:28:35.334496 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-5l6sj" Feb 02 15:28:35 crc kubenswrapper[4733]: I0202 15:28:35.997149 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-7pnkj" Feb 02 15:28:38 crc kubenswrapper[4733]: I0202 15:28:38.056004 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-c5q6r" Feb 02 15:28:38 crc kubenswrapper[4733]: I0202 15:28:38.107224 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-c5q6r"] Feb 02 15:28:38 crc kubenswrapper[4733]: I0202 15:28:38.559228 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kn8hn" Feb 02 15:28:38 crc kubenswrapper[4733]: I0202 15:28:38.560328 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kn8hn" Feb 02 15:28:38 crc kubenswrapper[4733]: I0202 15:28:38.627342 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-kn8hn" Feb 02 15:28:38 crc kubenswrapper[4733]: I0202 15:28:38.992804 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-c5q6r" podUID="3c3556a8-0e08-4f02-91cd-3b324e8dcbe8" containerName="registry-server" containerID="cri-o://d9dbc639cf53a165aa8dfdf5f9be047ba16af94c682cb64eae180ac3dbee03a8" gracePeriod=2 Feb 02 15:28:39 crc kubenswrapper[4733]: I0202 15:28:39.040613 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-kn8hn" Feb 02 15:28:39 crc kubenswrapper[4733]: I0202 15:28:39.442351 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c5q6r" Feb 02 15:28:39 crc kubenswrapper[4733]: I0202 15:28:39.589564 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c3556a8-0e08-4f02-91cd-3b324e8dcbe8-utilities\") pod \"3c3556a8-0e08-4f02-91cd-3b324e8dcbe8\" (UID: \"3c3556a8-0e08-4f02-91cd-3b324e8dcbe8\") " Feb 02 15:28:39 crc kubenswrapper[4733]: I0202 15:28:39.589651 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c3556a8-0e08-4f02-91cd-3b324e8dcbe8-catalog-content\") pod \"3c3556a8-0e08-4f02-91cd-3b324e8dcbe8\" (UID: \"3c3556a8-0e08-4f02-91cd-3b324e8dcbe8\") " Feb 02 15:28:39 crc kubenswrapper[4733]: I0202 15:28:39.589733 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9qm4x\" (UniqueName: \"kubernetes.io/projected/3c3556a8-0e08-4f02-91cd-3b324e8dcbe8-kube-api-access-9qm4x\") pod \"3c3556a8-0e08-4f02-91cd-3b324e8dcbe8\" (UID: \"3c3556a8-0e08-4f02-91cd-3b324e8dcbe8\") " Feb 02 15:28:39 crc kubenswrapper[4733]: I0202 15:28:39.590527 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c3556a8-0e08-4f02-91cd-3b324e8dcbe8-utilities" (OuterVolumeSpecName: "utilities") pod "3c3556a8-0e08-4f02-91cd-3b324e8dcbe8" (UID: "3c3556a8-0e08-4f02-91cd-3b324e8dcbe8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:28:39 crc kubenswrapper[4733]: I0202 15:28:39.600344 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c3556a8-0e08-4f02-91cd-3b324e8dcbe8-kube-api-access-9qm4x" (OuterVolumeSpecName: "kube-api-access-9qm4x") pod "3c3556a8-0e08-4f02-91cd-3b324e8dcbe8" (UID: "3c3556a8-0e08-4f02-91cd-3b324e8dcbe8"). InnerVolumeSpecName "kube-api-access-9qm4x". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:28:39 crc kubenswrapper[4733]: I0202 15:28:39.649263 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c3556a8-0e08-4f02-91cd-3b324e8dcbe8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3c3556a8-0e08-4f02-91cd-3b324e8dcbe8" (UID: "3c3556a8-0e08-4f02-91cd-3b324e8dcbe8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:28:39 crc kubenswrapper[4733]: I0202 15:28:39.691414 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c3556a8-0e08-4f02-91cd-3b324e8dcbe8-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 15:28:39 crc kubenswrapper[4733]: I0202 15:28:39.691456 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c3556a8-0e08-4f02-91cd-3b324e8dcbe8-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 15:28:39 crc kubenswrapper[4733]: I0202 15:28:39.691471 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9qm4x\" (UniqueName: \"kubernetes.io/projected/3c3556a8-0e08-4f02-91cd-3b324e8dcbe8-kube-api-access-9qm4x\") on node \"crc\" DevicePath \"\"" Feb 02 15:28:39 crc kubenswrapper[4733]: I0202 15:28:39.896117 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kn8hn"] Feb 02 15:28:40 crc kubenswrapper[4733]: I0202 15:28:40.001876 4733 generic.go:334] "Generic (PLEG): container finished" podID="3c3556a8-0e08-4f02-91cd-3b324e8dcbe8" containerID="d9dbc639cf53a165aa8dfdf5f9be047ba16af94c682cb64eae180ac3dbee03a8" exitCode=0 Feb 02 15:28:40 crc kubenswrapper[4733]: I0202 15:28:40.001934 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c5q6r" Feb 02 15:28:40 crc kubenswrapper[4733]: I0202 15:28:40.001970 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c5q6r" event={"ID":"3c3556a8-0e08-4f02-91cd-3b324e8dcbe8","Type":"ContainerDied","Data":"d9dbc639cf53a165aa8dfdf5f9be047ba16af94c682cb64eae180ac3dbee03a8"} Feb 02 15:28:40 crc kubenswrapper[4733]: I0202 15:28:40.002030 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c5q6r" event={"ID":"3c3556a8-0e08-4f02-91cd-3b324e8dcbe8","Type":"ContainerDied","Data":"5f4c0f856fec2e9b21315fde66b40fa5baaf18d73faa1c6bfa433a303f2723e8"} Feb 02 15:28:40 crc kubenswrapper[4733]: I0202 15:28:40.002048 4733 scope.go:117] "RemoveContainer" containerID="d9dbc639cf53a165aa8dfdf5f9be047ba16af94c682cb64eae180ac3dbee03a8" Feb 02 15:28:40 crc kubenswrapper[4733]: I0202 15:28:40.020418 4733 scope.go:117] "RemoveContainer" containerID="cf999c48149fa874d4f30fdb43f55656b7477f1478f5ff0263514c147101f7bb" Feb 02 15:28:40 crc kubenswrapper[4733]: I0202 15:28:40.041207 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-c5q6r"] Feb 02 15:28:40 crc kubenswrapper[4733]: I0202 15:28:40.044563 4733 scope.go:117] "RemoveContainer" containerID="2cf1e253ed16db53041b1c8290ee4c837ef70cedade2f6511a598e9efa417409" Feb 02 15:28:40 crc kubenswrapper[4733]: I0202 15:28:40.046120 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-c5q6r"] Feb 02 15:28:40 crc kubenswrapper[4733]: I0202 15:28:40.070307 4733 scope.go:117] "RemoveContainer" containerID="d9dbc639cf53a165aa8dfdf5f9be047ba16af94c682cb64eae180ac3dbee03a8" Feb 02 15:28:40 crc kubenswrapper[4733]: E0202 15:28:40.071778 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9dbc639cf53a165aa8dfdf5f9be047ba16af94c682cb64eae180ac3dbee03a8\": container with ID starting with d9dbc639cf53a165aa8dfdf5f9be047ba16af94c682cb64eae180ac3dbee03a8 not found: ID does not exist" containerID="d9dbc639cf53a165aa8dfdf5f9be047ba16af94c682cb64eae180ac3dbee03a8" Feb 02 15:28:40 crc kubenswrapper[4733]: I0202 15:28:40.071811 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9dbc639cf53a165aa8dfdf5f9be047ba16af94c682cb64eae180ac3dbee03a8"} err="failed to get container status \"d9dbc639cf53a165aa8dfdf5f9be047ba16af94c682cb64eae180ac3dbee03a8\": rpc error: code = NotFound desc = could not find container \"d9dbc639cf53a165aa8dfdf5f9be047ba16af94c682cb64eae180ac3dbee03a8\": container with ID starting with d9dbc639cf53a165aa8dfdf5f9be047ba16af94c682cb64eae180ac3dbee03a8 not found: ID does not exist" Feb 02 15:28:40 crc kubenswrapper[4733]: I0202 15:28:40.071835 4733 scope.go:117] "RemoveContainer" containerID="cf999c48149fa874d4f30fdb43f55656b7477f1478f5ff0263514c147101f7bb" Feb 02 15:28:40 crc kubenswrapper[4733]: E0202 15:28:40.072590 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf999c48149fa874d4f30fdb43f55656b7477f1478f5ff0263514c147101f7bb\": container with ID starting with cf999c48149fa874d4f30fdb43f55656b7477f1478f5ff0263514c147101f7bb not found: ID does not exist" containerID="cf999c48149fa874d4f30fdb43f55656b7477f1478f5ff0263514c147101f7bb" Feb 02 15:28:40 crc kubenswrapper[4733]: I0202 15:28:40.072653 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf999c48149fa874d4f30fdb43f55656b7477f1478f5ff0263514c147101f7bb"} err="failed to get container status \"cf999c48149fa874d4f30fdb43f55656b7477f1478f5ff0263514c147101f7bb\": rpc error: code = NotFound desc = could not find container \"cf999c48149fa874d4f30fdb43f55656b7477f1478f5ff0263514c147101f7bb\": container with ID starting with cf999c48149fa874d4f30fdb43f55656b7477f1478f5ff0263514c147101f7bb not found: ID does not exist" Feb 02 15:28:40 crc kubenswrapper[4733]: I0202 15:28:40.072696 4733 scope.go:117] "RemoveContainer" containerID="2cf1e253ed16db53041b1c8290ee4c837ef70cedade2f6511a598e9efa417409" Feb 02 15:28:40 crc kubenswrapper[4733]: E0202 15:28:40.073260 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2cf1e253ed16db53041b1c8290ee4c837ef70cedade2f6511a598e9efa417409\": container with ID starting with 2cf1e253ed16db53041b1c8290ee4c837ef70cedade2f6511a598e9efa417409 not found: ID does not exist" containerID="2cf1e253ed16db53041b1c8290ee4c837ef70cedade2f6511a598e9efa417409" Feb 02 15:28:40 crc kubenswrapper[4733]: I0202 15:28:40.073300 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2cf1e253ed16db53041b1c8290ee4c837ef70cedade2f6511a598e9efa417409"} err="failed to get container status \"2cf1e253ed16db53041b1c8290ee4c837ef70cedade2f6511a598e9efa417409\": rpc error: code = NotFound desc = could not find container \"2cf1e253ed16db53041b1c8290ee4c837ef70cedade2f6511a598e9efa417409\": container with ID starting with 2cf1e253ed16db53041b1c8290ee4c837ef70cedade2f6511a598e9efa417409 not found: ID does not exist" Feb 02 15:28:41 crc kubenswrapper[4733]: I0202 15:28:41.009972 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-kn8hn" podUID="afb40853-b770-44fb-a848-d23ea55ef184" containerName="registry-server" containerID="cri-o://b638997e132de496c2331dc3e264df17c6ba3e287577ba916d0dfd55c30417e5" gracePeriod=2 Feb 02 15:28:41 crc kubenswrapper[4733]: I0202 15:28:41.285587 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c3556a8-0e08-4f02-91cd-3b324e8dcbe8" path="/var/lib/kubelet/pods/3c3556a8-0e08-4f02-91cd-3b324e8dcbe8/volumes" Feb 02 15:28:41 crc kubenswrapper[4733]: I0202 15:28:41.430862 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kn8hn" Feb 02 15:28:41 crc kubenswrapper[4733]: I0202 15:28:41.615257 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-chvlq\" (UniqueName: \"kubernetes.io/projected/afb40853-b770-44fb-a848-d23ea55ef184-kube-api-access-chvlq\") pod \"afb40853-b770-44fb-a848-d23ea55ef184\" (UID: \"afb40853-b770-44fb-a848-d23ea55ef184\") " Feb 02 15:28:41 crc kubenswrapper[4733]: I0202 15:28:41.615374 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afb40853-b770-44fb-a848-d23ea55ef184-utilities\") pod \"afb40853-b770-44fb-a848-d23ea55ef184\" (UID: \"afb40853-b770-44fb-a848-d23ea55ef184\") " Feb 02 15:28:41 crc kubenswrapper[4733]: I0202 15:28:41.615443 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afb40853-b770-44fb-a848-d23ea55ef184-catalog-content\") pod \"afb40853-b770-44fb-a848-d23ea55ef184\" (UID: \"afb40853-b770-44fb-a848-d23ea55ef184\") " Feb 02 15:28:41 crc kubenswrapper[4733]: I0202 15:28:41.616474 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/afb40853-b770-44fb-a848-d23ea55ef184-utilities" (OuterVolumeSpecName: "utilities") pod "afb40853-b770-44fb-a848-d23ea55ef184" (UID: "afb40853-b770-44fb-a848-d23ea55ef184"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:28:41 crc kubenswrapper[4733]: I0202 15:28:41.624813 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afb40853-b770-44fb-a848-d23ea55ef184-kube-api-access-chvlq" (OuterVolumeSpecName: "kube-api-access-chvlq") pod "afb40853-b770-44fb-a848-d23ea55ef184" (UID: "afb40853-b770-44fb-a848-d23ea55ef184"). InnerVolumeSpecName "kube-api-access-chvlq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:28:41 crc kubenswrapper[4733]: I0202 15:28:41.653599 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/afb40853-b770-44fb-a848-d23ea55ef184-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "afb40853-b770-44fb-a848-d23ea55ef184" (UID: "afb40853-b770-44fb-a848-d23ea55ef184"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:28:41 crc kubenswrapper[4733]: I0202 15:28:41.716970 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afb40853-b770-44fb-a848-d23ea55ef184-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 15:28:41 crc kubenswrapper[4733]: I0202 15:28:41.717025 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afb40853-b770-44fb-a848-d23ea55ef184-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 15:28:41 crc kubenswrapper[4733]: I0202 15:28:41.717039 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-chvlq\" (UniqueName: \"kubernetes.io/projected/afb40853-b770-44fb-a848-d23ea55ef184-kube-api-access-chvlq\") on node \"crc\" DevicePath \"\"" Feb 02 15:28:42 crc kubenswrapper[4733]: I0202 15:28:42.020064 4733 generic.go:334] "Generic (PLEG): container finished" podID="afb40853-b770-44fb-a848-d23ea55ef184" containerID="b638997e132de496c2331dc3e264df17c6ba3e287577ba916d0dfd55c30417e5" exitCode=0 Feb 02 15:28:42 crc kubenswrapper[4733]: I0202 15:28:42.020114 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kn8hn" event={"ID":"afb40853-b770-44fb-a848-d23ea55ef184","Type":"ContainerDied","Data":"b638997e132de496c2331dc3e264df17c6ba3e287577ba916d0dfd55c30417e5"} Feb 02 15:28:42 crc kubenswrapper[4733]: I0202 15:28:42.020217 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kn8hn" Feb 02 15:28:42 crc kubenswrapper[4733]: I0202 15:28:42.020252 4733 scope.go:117] "RemoveContainer" containerID="b638997e132de496c2331dc3e264df17c6ba3e287577ba916d0dfd55c30417e5" Feb 02 15:28:42 crc kubenswrapper[4733]: I0202 15:28:42.020233 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kn8hn" event={"ID":"afb40853-b770-44fb-a848-d23ea55ef184","Type":"ContainerDied","Data":"bc96c332030f5f9655fbd440f2b8414bfd2c42cf55f6ebc5d1934f427ef418c2"} Feb 02 15:28:42 crc kubenswrapper[4733]: I0202 15:28:42.035952 4733 scope.go:117] "RemoveContainer" containerID="39f1449c1004537b0605cd0c1c497cd15c0936cef5c4bc3fd603b8abb6f9a684" Feb 02 15:28:42 crc kubenswrapper[4733]: I0202 15:28:42.051690 4733 scope.go:117] "RemoveContainer" containerID="7aae18ef19919774f2cde8911e5a4460acd630dac3077ffb2d26ad1ede6189d3" Feb 02 15:28:42 crc kubenswrapper[4733]: I0202 15:28:42.074690 4733 scope.go:117] "RemoveContainer" containerID="b638997e132de496c2331dc3e264df17c6ba3e287577ba916d0dfd55c30417e5" Feb 02 15:28:42 crc kubenswrapper[4733]: E0202 15:28:42.075367 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b638997e132de496c2331dc3e264df17c6ba3e287577ba916d0dfd55c30417e5\": container with ID starting with b638997e132de496c2331dc3e264df17c6ba3e287577ba916d0dfd55c30417e5 not found: ID does not exist" containerID="b638997e132de496c2331dc3e264df17c6ba3e287577ba916d0dfd55c30417e5" Feb 02 15:28:42 crc kubenswrapper[4733]: I0202 15:28:42.075412 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b638997e132de496c2331dc3e264df17c6ba3e287577ba916d0dfd55c30417e5"} err="failed to get container status \"b638997e132de496c2331dc3e264df17c6ba3e287577ba916d0dfd55c30417e5\": rpc error: code = NotFound desc = could not find container \"b638997e132de496c2331dc3e264df17c6ba3e287577ba916d0dfd55c30417e5\": container with ID starting with b638997e132de496c2331dc3e264df17c6ba3e287577ba916d0dfd55c30417e5 not found: ID does not exist" Feb 02 15:28:42 crc kubenswrapper[4733]: I0202 15:28:42.075443 4733 scope.go:117] "RemoveContainer" containerID="39f1449c1004537b0605cd0c1c497cd15c0936cef5c4bc3fd603b8abb6f9a684" Feb 02 15:28:42 crc kubenswrapper[4733]: E0202 15:28:42.075813 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39f1449c1004537b0605cd0c1c497cd15c0936cef5c4bc3fd603b8abb6f9a684\": container with ID starting with 39f1449c1004537b0605cd0c1c497cd15c0936cef5c4bc3fd603b8abb6f9a684 not found: ID does not exist" containerID="39f1449c1004537b0605cd0c1c497cd15c0936cef5c4bc3fd603b8abb6f9a684" Feb 02 15:28:42 crc kubenswrapper[4733]: I0202 15:28:42.075876 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39f1449c1004537b0605cd0c1c497cd15c0936cef5c4bc3fd603b8abb6f9a684"} err="failed to get container status \"39f1449c1004537b0605cd0c1c497cd15c0936cef5c4bc3fd603b8abb6f9a684\": rpc error: code = NotFound desc = could not find container \"39f1449c1004537b0605cd0c1c497cd15c0936cef5c4bc3fd603b8abb6f9a684\": container with ID starting with 39f1449c1004537b0605cd0c1c497cd15c0936cef5c4bc3fd603b8abb6f9a684 not found: ID does not exist" Feb 02 15:28:42 crc kubenswrapper[4733]: I0202 15:28:42.075911 4733 scope.go:117] "RemoveContainer" containerID="7aae18ef19919774f2cde8911e5a4460acd630dac3077ffb2d26ad1ede6189d3" Feb 02 15:28:42 crc kubenswrapper[4733]: E0202 15:28:42.076222 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7aae18ef19919774f2cde8911e5a4460acd630dac3077ffb2d26ad1ede6189d3\": container with ID starting with 7aae18ef19919774f2cde8911e5a4460acd630dac3077ffb2d26ad1ede6189d3 not found: ID does not exist" containerID="7aae18ef19919774f2cde8911e5a4460acd630dac3077ffb2d26ad1ede6189d3" Feb 02 15:28:42 crc kubenswrapper[4733]: I0202 15:28:42.076255 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7aae18ef19919774f2cde8911e5a4460acd630dac3077ffb2d26ad1ede6189d3"} err="failed to get container status \"7aae18ef19919774f2cde8911e5a4460acd630dac3077ffb2d26ad1ede6189d3\": rpc error: code = NotFound desc = could not find container \"7aae18ef19919774f2cde8911e5a4460acd630dac3077ffb2d26ad1ede6189d3\": container with ID starting with 7aae18ef19919774f2cde8911e5a4460acd630dac3077ffb2d26ad1ede6189d3 not found: ID does not exist" Feb 02 15:28:42 crc kubenswrapper[4733]: I0202 15:28:42.076514 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kn8hn"] Feb 02 15:28:42 crc kubenswrapper[4733]: I0202 15:28:42.086220 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-kn8hn"] Feb 02 15:28:43 crc kubenswrapper[4733]: I0202 15:28:43.263692 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="afb40853-b770-44fb-a848-d23ea55ef184" path="/var/lib/kubelet/pods/afb40853-b770-44fb-a848-d23ea55ef184/volumes" Feb 02 15:28:45 crc kubenswrapper[4733]: I0202 15:28:45.912753 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-index-bmfgc"] Feb 02 15:28:45 crc kubenswrapper[4733]: E0202 15:28:45.913674 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c3556a8-0e08-4f02-91cd-3b324e8dcbe8" containerName="extract-content" Feb 02 15:28:45 crc kubenswrapper[4733]: I0202 15:28:45.913702 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c3556a8-0e08-4f02-91cd-3b324e8dcbe8" containerName="extract-content" Feb 02 15:28:45 crc kubenswrapper[4733]: E0202 15:28:45.913723 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afb40853-b770-44fb-a848-d23ea55ef184" containerName="extract-content" Feb 02 15:28:45 crc kubenswrapper[4733]: I0202 15:28:45.913738 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="afb40853-b770-44fb-a848-d23ea55ef184" containerName="extract-content" Feb 02 15:28:45 crc kubenswrapper[4733]: E0202 15:28:45.913766 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afb40853-b770-44fb-a848-d23ea55ef184" containerName="registry-server" Feb 02 15:28:45 crc kubenswrapper[4733]: I0202 15:28:45.913784 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="afb40853-b770-44fb-a848-d23ea55ef184" containerName="registry-server" Feb 02 15:28:45 crc kubenswrapper[4733]: E0202 15:28:45.913924 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c3556a8-0e08-4f02-91cd-3b324e8dcbe8" containerName="registry-server" Feb 02 15:28:45 crc kubenswrapper[4733]: I0202 15:28:45.913990 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c3556a8-0e08-4f02-91cd-3b324e8dcbe8" containerName="registry-server" Feb 02 15:28:45 crc kubenswrapper[4733]: E0202 15:28:45.914004 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afb40853-b770-44fb-a848-d23ea55ef184" containerName="extract-utilities" Feb 02 15:28:45 crc kubenswrapper[4733]: I0202 15:28:45.914014 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="afb40853-b770-44fb-a848-d23ea55ef184" containerName="extract-utilities" Feb 02 15:28:45 crc kubenswrapper[4733]: E0202 15:28:45.914030 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c3556a8-0e08-4f02-91cd-3b324e8dcbe8" containerName="extract-utilities" Feb 02 15:28:45 crc kubenswrapper[4733]: I0202 15:28:45.914039 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c3556a8-0e08-4f02-91cd-3b324e8dcbe8" containerName="extract-utilities" Feb 02 15:28:45 crc kubenswrapper[4733]: I0202 15:28:45.914467 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="afb40853-b770-44fb-a848-d23ea55ef184" containerName="registry-server" Feb 02 15:28:45 crc kubenswrapper[4733]: I0202 15:28:45.914494 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c3556a8-0e08-4f02-91cd-3b324e8dcbe8" containerName="registry-server" Feb 02 15:28:45 crc kubenswrapper[4733]: I0202 15:28:45.915264 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-bmfgc" Feb 02 15:28:45 crc kubenswrapper[4733]: I0202 15:28:45.918980 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Feb 02 15:28:45 crc kubenswrapper[4733]: I0202 15:28:45.920400 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-index-dockercfg-tw878" Feb 02 15:28:45 crc kubenswrapper[4733]: I0202 15:28:45.920913 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Feb 02 15:28:45 crc kubenswrapper[4733]: I0202 15:28:45.930412 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-bmfgc"] Feb 02 15:28:46 crc kubenswrapper[4733]: I0202 15:28:46.073414 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p52nr\" (UniqueName: \"kubernetes.io/projected/1904d4f7-5ff6-4b7f-ba4a-5bf63ac6f95a-kube-api-access-p52nr\") pod \"mariadb-operator-index-bmfgc\" (UID: \"1904d4f7-5ff6-4b7f-ba4a-5bf63ac6f95a\") " pod="openstack-operators/mariadb-operator-index-bmfgc" Feb 02 15:28:46 crc kubenswrapper[4733]: I0202 15:28:46.175328 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p52nr\" (UniqueName: \"kubernetes.io/projected/1904d4f7-5ff6-4b7f-ba4a-5bf63ac6f95a-kube-api-access-p52nr\") pod \"mariadb-operator-index-bmfgc\" (UID: \"1904d4f7-5ff6-4b7f-ba4a-5bf63ac6f95a\") " pod="openstack-operators/mariadb-operator-index-bmfgc" Feb 02 15:28:46 crc kubenswrapper[4733]: I0202 15:28:46.208808 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p52nr\" (UniqueName: \"kubernetes.io/projected/1904d4f7-5ff6-4b7f-ba4a-5bf63ac6f95a-kube-api-access-p52nr\") pod \"mariadb-operator-index-bmfgc\" (UID: \"1904d4f7-5ff6-4b7f-ba4a-5bf63ac6f95a\") " pod="openstack-operators/mariadb-operator-index-bmfgc" Feb 02 15:28:46 crc kubenswrapper[4733]: I0202 15:28:46.236784 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-bmfgc" Feb 02 15:28:46 crc kubenswrapper[4733]: I0202 15:28:46.527256 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-bmfgc"] Feb 02 15:28:46 crc kubenswrapper[4733]: W0202 15:28:46.536951 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1904d4f7_5ff6_4b7f_ba4a_5bf63ac6f95a.slice/crio-e01105b3fbedb424c8f2df21c648539b996f8a48a3e431902bde68d2aeb1e749 WatchSource:0}: Error finding container e01105b3fbedb424c8f2df21c648539b996f8a48a3e431902bde68d2aeb1e749: Status 404 returned error can't find the container with id e01105b3fbedb424c8f2df21c648539b996f8a48a3e431902bde68d2aeb1e749 Feb 02 15:28:47 crc kubenswrapper[4733]: I0202 15:28:47.059328 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-bmfgc" event={"ID":"1904d4f7-5ff6-4b7f-ba4a-5bf63ac6f95a","Type":"ContainerStarted","Data":"e01105b3fbedb424c8f2df21c648539b996f8a48a3e431902bde68d2aeb1e749"} Feb 02 15:28:48 crc kubenswrapper[4733]: I0202 15:28:48.066444 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-bmfgc" event={"ID":"1904d4f7-5ff6-4b7f-ba4a-5bf63ac6f95a","Type":"ContainerStarted","Data":"b65578e7d7a4500ed6660821b27c40d43d0cc1fe5e3a180af3050acea260f47b"} Feb 02 15:28:48 crc kubenswrapper[4733]: I0202 15:28:48.088474 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-index-bmfgc" podStartSLOduration=2.102209572 podStartE2EDuration="3.08845532s" podCreationTimestamp="2026-02-02 15:28:45 +0000 UTC" firstStartedPulling="2026-02-02 15:28:46.539559335 +0000 UTC m=+809.991020693" lastFinishedPulling="2026-02-02 15:28:47.525805083 +0000 UTC m=+810.977266441" observedRunningTime="2026-02-02 15:28:48.085861863 +0000 UTC m=+811.537323231" watchObservedRunningTime="2026-02-02 15:28:48.08845532 +0000 UTC m=+811.539916688" Feb 02 15:28:56 crc kubenswrapper[4733]: I0202 15:28:56.237454 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/mariadb-operator-index-bmfgc" Feb 02 15:28:56 crc kubenswrapper[4733]: I0202 15:28:56.238209 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-index-bmfgc" Feb 02 15:28:56 crc kubenswrapper[4733]: I0202 15:28:56.283221 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/mariadb-operator-index-bmfgc" Feb 02 15:28:57 crc kubenswrapper[4733]: I0202 15:28:57.167822 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-index-bmfgc" Feb 02 15:28:58 crc kubenswrapper[4733]: I0202 15:28:58.760269 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40qpth5"] Feb 02 15:28:58 crc kubenswrapper[4733]: I0202 15:28:58.762324 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40qpth5" Feb 02 15:28:58 crc kubenswrapper[4733]: I0202 15:28:58.766117 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-2mtp8" Feb 02 15:28:58 crc kubenswrapper[4733]: I0202 15:28:58.782032 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40qpth5"] Feb 02 15:28:58 crc kubenswrapper[4733]: I0202 15:28:58.908746 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/40e65f33-ed0b-4d65-8a08-4a4ee970ea0f-util\") pod \"f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40qpth5\" (UID: \"40e65f33-ed0b-4d65-8a08-4a4ee970ea0f\") " pod="openstack-operators/f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40qpth5" Feb 02 15:28:58 crc kubenswrapper[4733]: I0202 15:28:58.908816 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxbnb\" (UniqueName: \"kubernetes.io/projected/40e65f33-ed0b-4d65-8a08-4a4ee970ea0f-kube-api-access-vxbnb\") pod \"f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40qpth5\" (UID: \"40e65f33-ed0b-4d65-8a08-4a4ee970ea0f\") " pod="openstack-operators/f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40qpth5" Feb 02 15:28:58 crc kubenswrapper[4733]: I0202 15:28:58.908919 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/40e65f33-ed0b-4d65-8a08-4a4ee970ea0f-bundle\") pod \"f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40qpth5\" (UID: \"40e65f33-ed0b-4d65-8a08-4a4ee970ea0f\") " pod="openstack-operators/f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40qpth5" Feb 02 15:28:59 crc kubenswrapper[4733]: I0202 15:28:59.009946 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxbnb\" (UniqueName: \"kubernetes.io/projected/40e65f33-ed0b-4d65-8a08-4a4ee970ea0f-kube-api-access-vxbnb\") pod \"f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40qpth5\" (UID: \"40e65f33-ed0b-4d65-8a08-4a4ee970ea0f\") " pod="openstack-operators/f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40qpth5" Feb 02 15:28:59 crc kubenswrapper[4733]: I0202 15:28:59.010062 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/40e65f33-ed0b-4d65-8a08-4a4ee970ea0f-bundle\") pod \"f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40qpth5\" (UID: \"40e65f33-ed0b-4d65-8a08-4a4ee970ea0f\") " pod="openstack-operators/f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40qpth5" Feb 02 15:28:59 crc kubenswrapper[4733]: I0202 15:28:59.010113 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/40e65f33-ed0b-4d65-8a08-4a4ee970ea0f-util\") pod \"f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40qpth5\" (UID: \"40e65f33-ed0b-4d65-8a08-4a4ee970ea0f\") " pod="openstack-operators/f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40qpth5" Feb 02 15:28:59 crc kubenswrapper[4733]: I0202 15:28:59.010710 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/40e65f33-ed0b-4d65-8a08-4a4ee970ea0f-util\") pod \"f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40qpth5\" (UID: \"40e65f33-ed0b-4d65-8a08-4a4ee970ea0f\") " pod="openstack-operators/f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40qpth5" Feb 02 15:28:59 crc kubenswrapper[4733]: I0202 15:28:59.010987 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/40e65f33-ed0b-4d65-8a08-4a4ee970ea0f-bundle\") pod \"f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40qpth5\" (UID: \"40e65f33-ed0b-4d65-8a08-4a4ee970ea0f\") " pod="openstack-operators/f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40qpth5" Feb 02 15:28:59 crc kubenswrapper[4733]: I0202 15:28:59.038010 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxbnb\" (UniqueName: \"kubernetes.io/projected/40e65f33-ed0b-4d65-8a08-4a4ee970ea0f-kube-api-access-vxbnb\") pod \"f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40qpth5\" (UID: \"40e65f33-ed0b-4d65-8a08-4a4ee970ea0f\") " pod="openstack-operators/f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40qpth5" Feb 02 15:28:59 crc kubenswrapper[4733]: I0202 15:28:59.107835 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40qpth5" Feb 02 15:28:59 crc kubenswrapper[4733]: I0202 15:28:59.363203 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40qpth5"] Feb 02 15:28:59 crc kubenswrapper[4733]: W0202 15:28:59.368527 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod40e65f33_ed0b_4d65_8a08_4a4ee970ea0f.slice/crio-31077240ff28ecade3da828435d9b34394f6470c37c505b10e481b54811d2c05 WatchSource:0}: Error finding container 31077240ff28ecade3da828435d9b34394f6470c37c505b10e481b54811d2c05: Status 404 returned error can't find the container with id 31077240ff28ecade3da828435d9b34394f6470c37c505b10e481b54811d2c05 Feb 02 15:29:00 crc kubenswrapper[4733]: I0202 15:29:00.163233 4733 generic.go:334] "Generic (PLEG): container finished" podID="40e65f33-ed0b-4d65-8a08-4a4ee970ea0f" containerID="1d6342cba8f24fd454efd3c05383711f7bc7a5f9589dc436535e902f27915680" exitCode=0 Feb 02 15:29:00 crc kubenswrapper[4733]: I0202 15:29:00.163320 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40qpth5" event={"ID":"40e65f33-ed0b-4d65-8a08-4a4ee970ea0f","Type":"ContainerDied","Data":"1d6342cba8f24fd454efd3c05383711f7bc7a5f9589dc436535e902f27915680"} Feb 02 15:29:00 crc kubenswrapper[4733]: I0202 15:29:00.163365 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40qpth5" event={"ID":"40e65f33-ed0b-4d65-8a08-4a4ee970ea0f","Type":"ContainerStarted","Data":"31077240ff28ecade3da828435d9b34394f6470c37c505b10e481b54811d2c05"} Feb 02 15:29:01 crc kubenswrapper[4733]: I0202 15:29:01.178787 4733 generic.go:334] "Generic (PLEG): container finished" podID="40e65f33-ed0b-4d65-8a08-4a4ee970ea0f" containerID="2bacaed5ef8ccec41d0fbf7135bc788ffa9e877f1c64af132fe7982b651e1f67" exitCode=0 Feb 02 15:29:01 crc kubenswrapper[4733]: I0202 15:29:01.178913 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40qpth5" event={"ID":"40e65f33-ed0b-4d65-8a08-4a4ee970ea0f","Type":"ContainerDied","Data":"2bacaed5ef8ccec41d0fbf7135bc788ffa9e877f1c64af132fe7982b651e1f67"} Feb 02 15:29:02 crc kubenswrapper[4733]: I0202 15:29:02.191134 4733 generic.go:334] "Generic (PLEG): container finished" podID="40e65f33-ed0b-4d65-8a08-4a4ee970ea0f" containerID="6b9792c10e3b3d8117ac5087d0b243117b2689263cdeec866a2c3fe00d543847" exitCode=0 Feb 02 15:29:02 crc kubenswrapper[4733]: I0202 15:29:02.191213 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40qpth5" event={"ID":"40e65f33-ed0b-4d65-8a08-4a4ee970ea0f","Type":"ContainerDied","Data":"6b9792c10e3b3d8117ac5087d0b243117b2689263cdeec866a2c3fe00d543847"} Feb 02 15:29:03 crc kubenswrapper[4733]: I0202 15:29:03.485463 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40qpth5" Feb 02 15:29:03 crc kubenswrapper[4733]: I0202 15:29:03.673414 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxbnb\" (UniqueName: \"kubernetes.io/projected/40e65f33-ed0b-4d65-8a08-4a4ee970ea0f-kube-api-access-vxbnb\") pod \"40e65f33-ed0b-4d65-8a08-4a4ee970ea0f\" (UID: \"40e65f33-ed0b-4d65-8a08-4a4ee970ea0f\") " Feb 02 15:29:03 crc kubenswrapper[4733]: I0202 15:29:03.673499 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/40e65f33-ed0b-4d65-8a08-4a4ee970ea0f-util\") pod \"40e65f33-ed0b-4d65-8a08-4a4ee970ea0f\" (UID: \"40e65f33-ed0b-4d65-8a08-4a4ee970ea0f\") " Feb 02 15:29:03 crc kubenswrapper[4733]: I0202 15:29:03.673568 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/40e65f33-ed0b-4d65-8a08-4a4ee970ea0f-bundle\") pod \"40e65f33-ed0b-4d65-8a08-4a4ee970ea0f\" (UID: \"40e65f33-ed0b-4d65-8a08-4a4ee970ea0f\") " Feb 02 15:29:03 crc kubenswrapper[4733]: I0202 15:29:03.675491 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40e65f33-ed0b-4d65-8a08-4a4ee970ea0f-bundle" (OuterVolumeSpecName: "bundle") pod "40e65f33-ed0b-4d65-8a08-4a4ee970ea0f" (UID: "40e65f33-ed0b-4d65-8a08-4a4ee970ea0f"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:29:03 crc kubenswrapper[4733]: I0202 15:29:03.708451 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40e65f33-ed0b-4d65-8a08-4a4ee970ea0f-util" (OuterVolumeSpecName: "util") pod "40e65f33-ed0b-4d65-8a08-4a4ee970ea0f" (UID: "40e65f33-ed0b-4d65-8a08-4a4ee970ea0f"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:29:03 crc kubenswrapper[4733]: I0202 15:29:03.775929 4733 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/40e65f33-ed0b-4d65-8a08-4a4ee970ea0f-util\") on node \"crc\" DevicePath \"\"" Feb 02 15:29:03 crc kubenswrapper[4733]: I0202 15:29:03.775985 4733 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/40e65f33-ed0b-4d65-8a08-4a4ee970ea0f-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 15:29:03 crc kubenswrapper[4733]: I0202 15:29:03.790561 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40e65f33-ed0b-4d65-8a08-4a4ee970ea0f-kube-api-access-vxbnb" (OuterVolumeSpecName: "kube-api-access-vxbnb") pod "40e65f33-ed0b-4d65-8a08-4a4ee970ea0f" (UID: "40e65f33-ed0b-4d65-8a08-4a4ee970ea0f"). InnerVolumeSpecName "kube-api-access-vxbnb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:29:03 crc kubenswrapper[4733]: I0202 15:29:03.878072 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxbnb\" (UniqueName: \"kubernetes.io/projected/40e65f33-ed0b-4d65-8a08-4a4ee970ea0f-kube-api-access-vxbnb\") on node \"crc\" DevicePath \"\"" Feb 02 15:29:04 crc kubenswrapper[4733]: I0202 15:29:04.207917 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40qpth5" event={"ID":"40e65f33-ed0b-4d65-8a08-4a4ee970ea0f","Type":"ContainerDied","Data":"31077240ff28ecade3da828435d9b34394f6470c37c505b10e481b54811d2c05"} Feb 02 15:29:04 crc kubenswrapper[4733]: I0202 15:29:04.207972 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="31077240ff28ecade3da828435d9b34394f6470c37c505b10e481b54811d2c05" Feb 02 15:29:04 crc kubenswrapper[4733]: I0202 15:29:04.207975 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40qpth5" Feb 02 15:29:04 crc kubenswrapper[4733]: I0202 15:29:04.988311 4733 patch_prober.go:28] interesting pod/machine-config-daemon-zgsv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 15:29:04 crc kubenswrapper[4733]: I0202 15:29:04.988387 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 15:29:09 crc kubenswrapper[4733]: I0202 15:29:09.363818 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6984697d4d-5mvzr"] Feb 02 15:29:09 crc kubenswrapper[4733]: E0202 15:29:09.364628 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40e65f33-ed0b-4d65-8a08-4a4ee970ea0f" containerName="pull" Feb 02 15:29:09 crc kubenswrapper[4733]: I0202 15:29:09.364644 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="40e65f33-ed0b-4d65-8a08-4a4ee970ea0f" containerName="pull" Feb 02 15:29:09 crc kubenswrapper[4733]: E0202 15:29:09.364658 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40e65f33-ed0b-4d65-8a08-4a4ee970ea0f" containerName="util" Feb 02 15:29:09 crc kubenswrapper[4733]: I0202 15:29:09.364667 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="40e65f33-ed0b-4d65-8a08-4a4ee970ea0f" containerName="util" Feb 02 15:29:09 crc kubenswrapper[4733]: E0202 15:29:09.364684 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40e65f33-ed0b-4d65-8a08-4a4ee970ea0f" containerName="extract" Feb 02 15:29:09 crc kubenswrapper[4733]: I0202 15:29:09.364693 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="40e65f33-ed0b-4d65-8a08-4a4ee970ea0f" containerName="extract" Feb 02 15:29:09 crc kubenswrapper[4733]: I0202 15:29:09.364816 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="40e65f33-ed0b-4d65-8a08-4a4ee970ea0f" containerName="extract" Feb 02 15:29:09 crc kubenswrapper[4733]: I0202 15:29:09.365491 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6984697d4d-5mvzr" Feb 02 15:29:09 crc kubenswrapper[4733]: I0202 15:29:09.367347 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Feb 02 15:29:09 crc kubenswrapper[4733]: I0202 15:29:09.367722 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-service-cert" Feb 02 15:29:09 crc kubenswrapper[4733]: I0202 15:29:09.367895 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-v9kj8" Feb 02 15:29:09 crc kubenswrapper[4733]: I0202 15:29:09.382932 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6984697d4d-5mvzr"] Feb 02 15:29:09 crc kubenswrapper[4733]: I0202 15:29:09.498956 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ccaf63aa-d3fe-492e-8f45-0502e4f66cd3-apiservice-cert\") pod \"mariadb-operator-controller-manager-6984697d4d-5mvzr\" (UID: \"ccaf63aa-d3fe-492e-8f45-0502e4f66cd3\") " pod="openstack-operators/mariadb-operator-controller-manager-6984697d4d-5mvzr" Feb 02 15:29:09 crc kubenswrapper[4733]: I0202 15:29:09.499029 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmjsm\" (UniqueName: \"kubernetes.io/projected/ccaf63aa-d3fe-492e-8f45-0502e4f66cd3-kube-api-access-zmjsm\") pod \"mariadb-operator-controller-manager-6984697d4d-5mvzr\" (UID: \"ccaf63aa-d3fe-492e-8f45-0502e4f66cd3\") " pod="openstack-operators/mariadb-operator-controller-manager-6984697d4d-5mvzr" Feb 02 15:29:09 crc kubenswrapper[4733]: I0202 15:29:09.499122 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ccaf63aa-d3fe-492e-8f45-0502e4f66cd3-webhook-cert\") pod \"mariadb-operator-controller-manager-6984697d4d-5mvzr\" (UID: \"ccaf63aa-d3fe-492e-8f45-0502e4f66cd3\") " pod="openstack-operators/mariadb-operator-controller-manager-6984697d4d-5mvzr" Feb 02 15:29:09 crc kubenswrapper[4733]: I0202 15:29:09.600527 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ccaf63aa-d3fe-492e-8f45-0502e4f66cd3-webhook-cert\") pod \"mariadb-operator-controller-manager-6984697d4d-5mvzr\" (UID: \"ccaf63aa-d3fe-492e-8f45-0502e4f66cd3\") " pod="openstack-operators/mariadb-operator-controller-manager-6984697d4d-5mvzr" Feb 02 15:29:09 crc kubenswrapper[4733]: I0202 15:29:09.600584 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ccaf63aa-d3fe-492e-8f45-0502e4f66cd3-apiservice-cert\") pod \"mariadb-operator-controller-manager-6984697d4d-5mvzr\" (UID: \"ccaf63aa-d3fe-492e-8f45-0502e4f66cd3\") " pod="openstack-operators/mariadb-operator-controller-manager-6984697d4d-5mvzr" Feb 02 15:29:09 crc kubenswrapper[4733]: I0202 15:29:09.600617 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmjsm\" (UniqueName: \"kubernetes.io/projected/ccaf63aa-d3fe-492e-8f45-0502e4f66cd3-kube-api-access-zmjsm\") pod \"mariadb-operator-controller-manager-6984697d4d-5mvzr\" (UID: \"ccaf63aa-d3fe-492e-8f45-0502e4f66cd3\") " pod="openstack-operators/mariadb-operator-controller-manager-6984697d4d-5mvzr" Feb 02 15:29:09 crc kubenswrapper[4733]: I0202 15:29:09.606836 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ccaf63aa-d3fe-492e-8f45-0502e4f66cd3-webhook-cert\") pod \"mariadb-operator-controller-manager-6984697d4d-5mvzr\" (UID: \"ccaf63aa-d3fe-492e-8f45-0502e4f66cd3\") " pod="openstack-operators/mariadb-operator-controller-manager-6984697d4d-5mvzr" Feb 02 15:29:09 crc kubenswrapper[4733]: I0202 15:29:09.607151 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ccaf63aa-d3fe-492e-8f45-0502e4f66cd3-apiservice-cert\") pod \"mariadb-operator-controller-manager-6984697d4d-5mvzr\" (UID: \"ccaf63aa-d3fe-492e-8f45-0502e4f66cd3\") " pod="openstack-operators/mariadb-operator-controller-manager-6984697d4d-5mvzr" Feb 02 15:29:09 crc kubenswrapper[4733]: I0202 15:29:09.620723 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmjsm\" (UniqueName: \"kubernetes.io/projected/ccaf63aa-d3fe-492e-8f45-0502e4f66cd3-kube-api-access-zmjsm\") pod \"mariadb-operator-controller-manager-6984697d4d-5mvzr\" (UID: \"ccaf63aa-d3fe-492e-8f45-0502e4f66cd3\") " pod="openstack-operators/mariadb-operator-controller-manager-6984697d4d-5mvzr" Feb 02 15:29:09 crc kubenswrapper[4733]: I0202 15:29:09.682013 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6984697d4d-5mvzr" Feb 02 15:29:10 crc kubenswrapper[4733]: I0202 15:29:10.126717 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6984697d4d-5mvzr"] Feb 02 15:29:10 crc kubenswrapper[4733]: I0202 15:29:10.244645 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6984697d4d-5mvzr" event={"ID":"ccaf63aa-d3fe-492e-8f45-0502e4f66cd3","Type":"ContainerStarted","Data":"88ccd8fde787861154d902dde44cb3bcb6c57e268ef0ec80c9d4a70d9761f163"} Feb 02 15:29:14 crc kubenswrapper[4733]: I0202 15:29:14.271894 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6984697d4d-5mvzr" event={"ID":"ccaf63aa-d3fe-492e-8f45-0502e4f66cd3","Type":"ContainerStarted","Data":"fc36ca98de458bfe957316ab2c04717dbf8f5a235e9cca3517345554e008d530"} Feb 02 15:29:14 crc kubenswrapper[4733]: I0202 15:29:14.272608 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-6984697d4d-5mvzr" Feb 02 15:29:14 crc kubenswrapper[4733]: I0202 15:29:14.289739 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-6984697d4d-5mvzr" podStartSLOduration=1.520321165 podStartE2EDuration="5.289720915s" podCreationTimestamp="2026-02-02 15:29:09 +0000 UTC" firstStartedPulling="2026-02-02 15:29:10.1374987 +0000 UTC m=+833.588960048" lastFinishedPulling="2026-02-02 15:29:13.90689844 +0000 UTC m=+837.358359798" observedRunningTime="2026-02-02 15:29:14.289098329 +0000 UTC m=+837.740559697" watchObservedRunningTime="2026-02-02 15:29:14.289720915 +0000 UTC m=+837.741182283" Feb 02 15:29:19 crc kubenswrapper[4733]: I0202 15:29:19.687199 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-6984697d4d-5mvzr" Feb 02 15:29:26 crc kubenswrapper[4733]: I0202 15:29:26.305259 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-index-8xpw7"] Feb 02 15:29:26 crc kubenswrapper[4733]: I0202 15:29:26.307192 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-8xpw7" Feb 02 15:29:26 crc kubenswrapper[4733]: I0202 15:29:26.309325 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-index-dockercfg-gg7mq" Feb 02 15:29:26 crc kubenswrapper[4733]: I0202 15:29:26.310642 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-8xpw7"] Feb 02 15:29:26 crc kubenswrapper[4733]: I0202 15:29:26.470012 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q52jv\" (UniqueName: \"kubernetes.io/projected/74233ea3-1961-4292-b58f-a4efaac0dfae-kube-api-access-q52jv\") pod \"infra-operator-index-8xpw7\" (UID: \"74233ea3-1961-4292-b58f-a4efaac0dfae\") " pod="openstack-operators/infra-operator-index-8xpw7" Feb 02 15:29:26 crc kubenswrapper[4733]: I0202 15:29:26.570796 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q52jv\" (UniqueName: \"kubernetes.io/projected/74233ea3-1961-4292-b58f-a4efaac0dfae-kube-api-access-q52jv\") pod \"infra-operator-index-8xpw7\" (UID: \"74233ea3-1961-4292-b58f-a4efaac0dfae\") " pod="openstack-operators/infra-operator-index-8xpw7" Feb 02 15:29:26 crc kubenswrapper[4733]: I0202 15:29:26.604304 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q52jv\" (UniqueName: \"kubernetes.io/projected/74233ea3-1961-4292-b58f-a4efaac0dfae-kube-api-access-q52jv\") pod \"infra-operator-index-8xpw7\" (UID: \"74233ea3-1961-4292-b58f-a4efaac0dfae\") " pod="openstack-operators/infra-operator-index-8xpw7" Feb 02 15:29:26 crc kubenswrapper[4733]: I0202 15:29:26.638956 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-8xpw7" Feb 02 15:29:27 crc kubenswrapper[4733]: I0202 15:29:27.111664 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-8xpw7"] Feb 02 15:29:27 crc kubenswrapper[4733]: I0202 15:29:27.348083 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-8xpw7" event={"ID":"74233ea3-1961-4292-b58f-a4efaac0dfae","Type":"ContainerStarted","Data":"7dd9b95b83157d009979cc5620fe34f848e04d52ec5abc461cc86b69317261cd"} Feb 02 15:29:28 crc kubenswrapper[4733]: I0202 15:29:28.355961 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-8xpw7" event={"ID":"74233ea3-1961-4292-b58f-a4efaac0dfae","Type":"ContainerStarted","Data":"05df8f0d229cf267db868cb0fdd192dce5947e761c3082bb3c8cef2f96aa3263"} Feb 02 15:29:28 crc kubenswrapper[4733]: I0202 15:29:28.377559 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-index-8xpw7" podStartSLOduration=1.5462350900000001 podStartE2EDuration="2.377526555s" podCreationTimestamp="2026-02-02 15:29:26 +0000 UTC" firstStartedPulling="2026-02-02 15:29:27.119414725 +0000 UTC m=+850.570876093" lastFinishedPulling="2026-02-02 15:29:27.9507062 +0000 UTC m=+851.402167558" observedRunningTime="2026-02-02 15:29:28.373392087 +0000 UTC m=+851.824853455" watchObservedRunningTime="2026-02-02 15:29:28.377526555 +0000 UTC m=+851.828987923" Feb 02 15:29:30 crc kubenswrapper[4733]: I0202 15:29:30.494308 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-index-8xpw7"] Feb 02 15:29:30 crc kubenswrapper[4733]: I0202 15:29:30.494959 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/infra-operator-index-8xpw7" podUID="74233ea3-1961-4292-b58f-a4efaac0dfae" containerName="registry-server" containerID="cri-o://05df8f0d229cf267db868cb0fdd192dce5947e761c3082bb3c8cef2f96aa3263" gracePeriod=2 Feb 02 15:29:30 crc kubenswrapper[4733]: I0202 15:29:30.948883 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-8xpw7" Feb 02 15:29:31 crc kubenswrapper[4733]: I0202 15:29:31.096920 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-index-552jg"] Feb 02 15:29:31 crc kubenswrapper[4733]: E0202 15:29:31.097371 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74233ea3-1961-4292-b58f-a4efaac0dfae" containerName="registry-server" Feb 02 15:29:31 crc kubenswrapper[4733]: I0202 15:29:31.097453 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="74233ea3-1961-4292-b58f-a4efaac0dfae" containerName="registry-server" Feb 02 15:29:31 crc kubenswrapper[4733]: I0202 15:29:31.097684 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="74233ea3-1961-4292-b58f-a4efaac0dfae" containerName="registry-server" Feb 02 15:29:31 crc kubenswrapper[4733]: I0202 15:29:31.098183 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-552jg" Feb 02 15:29:31 crc kubenswrapper[4733]: I0202 15:29:31.107077 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-552jg"] Feb 02 15:29:31 crc kubenswrapper[4733]: I0202 15:29:31.133946 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q52jv\" (UniqueName: \"kubernetes.io/projected/74233ea3-1961-4292-b58f-a4efaac0dfae-kube-api-access-q52jv\") pod \"74233ea3-1961-4292-b58f-a4efaac0dfae\" (UID: \"74233ea3-1961-4292-b58f-a4efaac0dfae\") " Feb 02 15:29:31 crc kubenswrapper[4733]: I0202 15:29:31.140094 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74233ea3-1961-4292-b58f-a4efaac0dfae-kube-api-access-q52jv" (OuterVolumeSpecName: "kube-api-access-q52jv") pod "74233ea3-1961-4292-b58f-a4efaac0dfae" (UID: "74233ea3-1961-4292-b58f-a4efaac0dfae"). InnerVolumeSpecName "kube-api-access-q52jv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:29:31 crc kubenswrapper[4733]: I0202 15:29:31.235837 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mph99\" (UniqueName: \"kubernetes.io/projected/bf01001b-fb23-4f09-8f88-38bce44a93ce-kube-api-access-mph99\") pod \"infra-operator-index-552jg\" (UID: \"bf01001b-fb23-4f09-8f88-38bce44a93ce\") " pod="openstack-operators/infra-operator-index-552jg" Feb 02 15:29:31 crc kubenswrapper[4733]: I0202 15:29:31.236116 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q52jv\" (UniqueName: \"kubernetes.io/projected/74233ea3-1961-4292-b58f-a4efaac0dfae-kube-api-access-q52jv\") on node \"crc\" DevicePath \"\"" Feb 02 15:29:31 crc kubenswrapper[4733]: I0202 15:29:31.337295 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mph99\" (UniqueName: \"kubernetes.io/projected/bf01001b-fb23-4f09-8f88-38bce44a93ce-kube-api-access-mph99\") pod \"infra-operator-index-552jg\" (UID: \"bf01001b-fb23-4f09-8f88-38bce44a93ce\") " pod="openstack-operators/infra-operator-index-552jg" Feb 02 15:29:31 crc kubenswrapper[4733]: I0202 15:29:31.367852 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mph99\" (UniqueName: \"kubernetes.io/projected/bf01001b-fb23-4f09-8f88-38bce44a93ce-kube-api-access-mph99\") pod \"infra-operator-index-552jg\" (UID: \"bf01001b-fb23-4f09-8f88-38bce44a93ce\") " pod="openstack-operators/infra-operator-index-552jg" Feb 02 15:29:31 crc kubenswrapper[4733]: I0202 15:29:31.378262 4733 generic.go:334] "Generic (PLEG): container finished" podID="74233ea3-1961-4292-b58f-a4efaac0dfae" containerID="05df8f0d229cf267db868cb0fdd192dce5947e761c3082bb3c8cef2f96aa3263" exitCode=0 Feb 02 15:29:31 crc kubenswrapper[4733]: I0202 15:29:31.378321 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-8xpw7" event={"ID":"74233ea3-1961-4292-b58f-a4efaac0dfae","Type":"ContainerDied","Data":"05df8f0d229cf267db868cb0fdd192dce5947e761c3082bb3c8cef2f96aa3263"} Feb 02 15:29:31 crc kubenswrapper[4733]: I0202 15:29:31.378357 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-8xpw7" event={"ID":"74233ea3-1961-4292-b58f-a4efaac0dfae","Type":"ContainerDied","Data":"7dd9b95b83157d009979cc5620fe34f848e04d52ec5abc461cc86b69317261cd"} Feb 02 15:29:31 crc kubenswrapper[4733]: I0202 15:29:31.378392 4733 scope.go:117] "RemoveContainer" containerID="05df8f0d229cf267db868cb0fdd192dce5947e761c3082bb3c8cef2f96aa3263" Feb 02 15:29:31 crc kubenswrapper[4733]: I0202 15:29:31.378791 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-8xpw7" Feb 02 15:29:31 crc kubenswrapper[4733]: I0202 15:29:31.398804 4733 scope.go:117] "RemoveContainer" containerID="05df8f0d229cf267db868cb0fdd192dce5947e761c3082bb3c8cef2f96aa3263" Feb 02 15:29:31 crc kubenswrapper[4733]: E0202 15:29:31.399521 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05df8f0d229cf267db868cb0fdd192dce5947e761c3082bb3c8cef2f96aa3263\": container with ID starting with 05df8f0d229cf267db868cb0fdd192dce5947e761c3082bb3c8cef2f96aa3263 not found: ID does not exist" containerID="05df8f0d229cf267db868cb0fdd192dce5947e761c3082bb3c8cef2f96aa3263" Feb 02 15:29:31 crc kubenswrapper[4733]: I0202 15:29:31.399571 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05df8f0d229cf267db868cb0fdd192dce5947e761c3082bb3c8cef2f96aa3263"} err="failed to get container status \"05df8f0d229cf267db868cb0fdd192dce5947e761c3082bb3c8cef2f96aa3263\": rpc error: code = NotFound desc = could not find container \"05df8f0d229cf267db868cb0fdd192dce5947e761c3082bb3c8cef2f96aa3263\": container with ID starting with 05df8f0d229cf267db868cb0fdd192dce5947e761c3082bb3c8cef2f96aa3263 not found: ID does not exist" Feb 02 15:29:31 crc kubenswrapper[4733]: I0202 15:29:31.400185 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-index-8xpw7"] Feb 02 15:29:31 crc kubenswrapper[4733]: I0202 15:29:31.403491 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/infra-operator-index-8xpw7"] Feb 02 15:29:31 crc kubenswrapper[4733]: I0202 15:29:31.417196 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-552jg" Feb 02 15:29:31 crc kubenswrapper[4733]: I0202 15:29:31.938994 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-552jg"] Feb 02 15:29:31 crc kubenswrapper[4733]: W0202 15:29:31.955935 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbf01001b_fb23_4f09_8f88_38bce44a93ce.slice/crio-72e9a7b68560d3215b10f90b7817deb03c361d6c35ac7081762db112eca7da16 WatchSource:0}: Error finding container 72e9a7b68560d3215b10f90b7817deb03c361d6c35ac7081762db112eca7da16: Status 404 returned error can't find the container with id 72e9a7b68560d3215b10f90b7817deb03c361d6c35ac7081762db112eca7da16 Feb 02 15:29:32 crc kubenswrapper[4733]: I0202 15:29:32.386428 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-552jg" event={"ID":"bf01001b-fb23-4f09-8f88-38bce44a93ce","Type":"ContainerStarted","Data":"72e9a7b68560d3215b10f90b7817deb03c361d6c35ac7081762db112eca7da16"} Feb 02 15:29:33 crc kubenswrapper[4733]: I0202 15:29:33.268823 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74233ea3-1961-4292-b58f-a4efaac0dfae" path="/var/lib/kubelet/pods/74233ea3-1961-4292-b58f-a4efaac0dfae/volumes" Feb 02 15:29:33 crc kubenswrapper[4733]: I0202 15:29:33.395709 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-552jg" event={"ID":"bf01001b-fb23-4f09-8f88-38bce44a93ce","Type":"ContainerStarted","Data":"3368bbba9d3374e124d780779edd0479bbb6d0926fdb67ad3475b13e53ce6e61"} Feb 02 15:29:33 crc kubenswrapper[4733]: I0202 15:29:33.425459 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-index-552jg" podStartSLOduration=1.697349643 podStartE2EDuration="2.425436056s" podCreationTimestamp="2026-02-02 15:29:31 +0000 UTC" firstStartedPulling="2026-02-02 15:29:31.961447786 +0000 UTC m=+855.412909184" lastFinishedPulling="2026-02-02 15:29:32.689534199 +0000 UTC m=+856.140995597" observedRunningTime="2026-02-02 15:29:33.420510337 +0000 UTC m=+856.871971735" watchObservedRunningTime="2026-02-02 15:29:33.425436056 +0000 UTC m=+856.876897424" Feb 02 15:29:34 crc kubenswrapper[4733]: I0202 15:29:34.988285 4733 patch_prober.go:28] interesting pod/machine-config-daemon-zgsv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 15:29:34 crc kubenswrapper[4733]: I0202 15:29:34.988647 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 15:29:34 crc kubenswrapper[4733]: I0202 15:29:34.988707 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" Feb 02 15:29:34 crc kubenswrapper[4733]: I0202 15:29:34.989626 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"25cacf7f1a9153fe0cadb8062fa48df6b9242dcf030528fe627cf68ad889d6b1"} pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 15:29:34 crc kubenswrapper[4733]: I0202 15:29:34.989735 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" containerName="machine-config-daemon" containerID="cri-o://25cacf7f1a9153fe0cadb8062fa48df6b9242dcf030528fe627cf68ad889d6b1" gracePeriod=600 Feb 02 15:29:35 crc kubenswrapper[4733]: I0202 15:29:35.417084 4733 generic.go:334] "Generic (PLEG): container finished" podID="83d0e810-ea5c-47aa-947b-7865b045bf94" containerID="25cacf7f1a9153fe0cadb8062fa48df6b9242dcf030528fe627cf68ad889d6b1" exitCode=0 Feb 02 15:29:35 crc kubenswrapper[4733]: I0202 15:29:35.417229 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" event={"ID":"83d0e810-ea5c-47aa-947b-7865b045bf94","Type":"ContainerDied","Data":"25cacf7f1a9153fe0cadb8062fa48df6b9242dcf030528fe627cf68ad889d6b1"} Feb 02 15:29:35 crc kubenswrapper[4733]: I0202 15:29:35.417508 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" event={"ID":"83d0e810-ea5c-47aa-947b-7865b045bf94","Type":"ContainerStarted","Data":"b050108ccfab6b78f46eeabfccf3d93e55ea50151667d0fea583d0f4073f7662"} Feb 02 15:29:35 crc kubenswrapper[4733]: I0202 15:29:35.417537 4733 scope.go:117] "RemoveContainer" containerID="6cb02659ae49ba1eb2bfa677c715cc0aaf1abb6c83347c16358782946f3f3f0f" Feb 02 15:29:41 crc kubenswrapper[4733]: I0202 15:29:41.418036 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/infra-operator-index-552jg" Feb 02 15:29:41 crc kubenswrapper[4733]: I0202 15:29:41.418638 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-index-552jg" Feb 02 15:29:41 crc kubenswrapper[4733]: I0202 15:29:41.466204 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/infra-operator-index-552jg" Feb 02 15:29:41 crc kubenswrapper[4733]: I0202 15:29:41.507653 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-index-552jg" Feb 02 15:29:43 crc kubenswrapper[4733]: I0202 15:29:43.355378 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f75764trqm"] Feb 02 15:29:43 crc kubenswrapper[4733]: I0202 15:29:43.358067 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f75764trqm" Feb 02 15:29:43 crc kubenswrapper[4733]: I0202 15:29:43.361369 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-2mtp8" Feb 02 15:29:43 crc kubenswrapper[4733]: I0202 15:29:43.373364 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f75764trqm"] Feb 02 15:29:43 crc kubenswrapper[4733]: I0202 15:29:43.526899 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f107fa8c-f81f-4d3f-8a14-e718942491b8-bundle\") pod \"d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f75764trqm\" (UID: \"f107fa8c-f81f-4d3f-8a14-e718942491b8\") " pod="openstack-operators/d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f75764trqm" Feb 02 15:29:43 crc kubenswrapper[4733]: I0202 15:29:43.526979 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wltk7\" (UniqueName: \"kubernetes.io/projected/f107fa8c-f81f-4d3f-8a14-e718942491b8-kube-api-access-wltk7\") pod \"d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f75764trqm\" (UID: \"f107fa8c-f81f-4d3f-8a14-e718942491b8\") " pod="openstack-operators/d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f75764trqm" Feb 02 15:29:43 crc kubenswrapper[4733]: I0202 15:29:43.527027 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f107fa8c-f81f-4d3f-8a14-e718942491b8-util\") pod \"d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f75764trqm\" (UID: \"f107fa8c-f81f-4d3f-8a14-e718942491b8\") " pod="openstack-operators/d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f75764trqm" Feb 02 15:29:43 crc kubenswrapper[4733]: I0202 15:29:43.628324 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wltk7\" (UniqueName: \"kubernetes.io/projected/f107fa8c-f81f-4d3f-8a14-e718942491b8-kube-api-access-wltk7\") pod \"d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f75764trqm\" (UID: \"f107fa8c-f81f-4d3f-8a14-e718942491b8\") " pod="openstack-operators/d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f75764trqm" Feb 02 15:29:43 crc kubenswrapper[4733]: I0202 15:29:43.628408 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f107fa8c-f81f-4d3f-8a14-e718942491b8-util\") pod \"d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f75764trqm\" (UID: \"f107fa8c-f81f-4d3f-8a14-e718942491b8\") " pod="openstack-operators/d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f75764trqm" Feb 02 15:29:43 crc kubenswrapper[4733]: I0202 15:29:43.628587 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f107fa8c-f81f-4d3f-8a14-e718942491b8-bundle\") pod \"d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f75764trqm\" (UID: \"f107fa8c-f81f-4d3f-8a14-e718942491b8\") " pod="openstack-operators/d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f75764trqm" Feb 02 15:29:43 crc kubenswrapper[4733]: I0202 15:29:43.629021 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f107fa8c-f81f-4d3f-8a14-e718942491b8-util\") pod \"d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f75764trqm\" (UID: \"f107fa8c-f81f-4d3f-8a14-e718942491b8\") " pod="openstack-operators/d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f75764trqm" Feb 02 15:29:43 crc kubenswrapper[4733]: I0202 15:29:43.629239 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f107fa8c-f81f-4d3f-8a14-e718942491b8-bundle\") pod \"d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f75764trqm\" (UID: \"f107fa8c-f81f-4d3f-8a14-e718942491b8\") " pod="openstack-operators/d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f75764trqm" Feb 02 15:29:43 crc kubenswrapper[4733]: I0202 15:29:43.667705 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wltk7\" (UniqueName: \"kubernetes.io/projected/f107fa8c-f81f-4d3f-8a14-e718942491b8-kube-api-access-wltk7\") pod \"d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f75764trqm\" (UID: \"f107fa8c-f81f-4d3f-8a14-e718942491b8\") " pod="openstack-operators/d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f75764trqm" Feb 02 15:29:43 crc kubenswrapper[4733]: I0202 15:29:43.692022 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f75764trqm" Feb 02 15:29:43 crc kubenswrapper[4733]: I0202 15:29:43.949286 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f75764trqm"] Feb 02 15:29:44 crc kubenswrapper[4733]: I0202 15:29:44.490099 4733 generic.go:334] "Generic (PLEG): container finished" podID="f107fa8c-f81f-4d3f-8a14-e718942491b8" containerID="ab19f2f2d434db0186b91ba8763d53b716f7e1126e222060a31f15c1c270df31" exitCode=0 Feb 02 15:29:44 crc kubenswrapper[4733]: I0202 15:29:44.490247 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f75764trqm" event={"ID":"f107fa8c-f81f-4d3f-8a14-e718942491b8","Type":"ContainerDied","Data":"ab19f2f2d434db0186b91ba8763d53b716f7e1126e222060a31f15c1c270df31"} Feb 02 15:29:44 crc kubenswrapper[4733]: I0202 15:29:44.490568 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f75764trqm" event={"ID":"f107fa8c-f81f-4d3f-8a14-e718942491b8","Type":"ContainerStarted","Data":"d945384e860d5e02676c1fe38f2f3e835220dbceda859d2fd9295268a7255d09"} Feb 02 15:29:45 crc kubenswrapper[4733]: I0202 15:29:45.498721 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f75764trqm" event={"ID":"f107fa8c-f81f-4d3f-8a14-e718942491b8","Type":"ContainerStarted","Data":"77f034135db5431b38fb66fe9ca3c5c0eb6838124fe6c9d890a8238d3715408a"} Feb 02 15:29:46 crc kubenswrapper[4733]: I0202 15:29:46.510881 4733 generic.go:334] "Generic (PLEG): container finished" podID="f107fa8c-f81f-4d3f-8a14-e718942491b8" containerID="77f034135db5431b38fb66fe9ca3c5c0eb6838124fe6c9d890a8238d3715408a" exitCode=0 Feb 02 15:29:46 crc kubenswrapper[4733]: I0202 15:29:46.510969 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f75764trqm" event={"ID":"f107fa8c-f81f-4d3f-8a14-e718942491b8","Type":"ContainerDied","Data":"77f034135db5431b38fb66fe9ca3c5c0eb6838124fe6c9d890a8238d3715408a"} Feb 02 15:29:47 crc kubenswrapper[4733]: I0202 15:29:47.523587 4733 generic.go:334] "Generic (PLEG): container finished" podID="f107fa8c-f81f-4d3f-8a14-e718942491b8" containerID="a95fee8d9d9b365d3b0c6597d3f5027e679d594c5a1448b4a40af3af0022b2f2" exitCode=0 Feb 02 15:29:47 crc kubenswrapper[4733]: I0202 15:29:47.523701 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f75764trqm" event={"ID":"f107fa8c-f81f-4d3f-8a14-e718942491b8","Type":"ContainerDied","Data":"a95fee8d9d9b365d3b0c6597d3f5027e679d594c5a1448b4a40af3af0022b2f2"} Feb 02 15:29:48 crc kubenswrapper[4733]: I0202 15:29:48.778245 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f75764trqm" Feb 02 15:29:48 crc kubenswrapper[4733]: I0202 15:29:48.907205 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f107fa8c-f81f-4d3f-8a14-e718942491b8-bundle\") pod \"f107fa8c-f81f-4d3f-8a14-e718942491b8\" (UID: \"f107fa8c-f81f-4d3f-8a14-e718942491b8\") " Feb 02 15:29:48 crc kubenswrapper[4733]: I0202 15:29:48.907294 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wltk7\" (UniqueName: \"kubernetes.io/projected/f107fa8c-f81f-4d3f-8a14-e718942491b8-kube-api-access-wltk7\") pod \"f107fa8c-f81f-4d3f-8a14-e718942491b8\" (UID: \"f107fa8c-f81f-4d3f-8a14-e718942491b8\") " Feb 02 15:29:48 crc kubenswrapper[4733]: I0202 15:29:48.907345 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f107fa8c-f81f-4d3f-8a14-e718942491b8-util\") pod \"f107fa8c-f81f-4d3f-8a14-e718942491b8\" (UID: \"f107fa8c-f81f-4d3f-8a14-e718942491b8\") " Feb 02 15:29:48 crc kubenswrapper[4733]: I0202 15:29:48.910470 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f107fa8c-f81f-4d3f-8a14-e718942491b8-bundle" (OuterVolumeSpecName: "bundle") pod "f107fa8c-f81f-4d3f-8a14-e718942491b8" (UID: "f107fa8c-f81f-4d3f-8a14-e718942491b8"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:29:48 crc kubenswrapper[4733]: I0202 15:29:48.916664 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f107fa8c-f81f-4d3f-8a14-e718942491b8-kube-api-access-wltk7" (OuterVolumeSpecName: "kube-api-access-wltk7") pod "f107fa8c-f81f-4d3f-8a14-e718942491b8" (UID: "f107fa8c-f81f-4d3f-8a14-e718942491b8"). InnerVolumeSpecName "kube-api-access-wltk7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:29:48 crc kubenswrapper[4733]: I0202 15:29:48.942969 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f107fa8c-f81f-4d3f-8a14-e718942491b8-util" (OuterVolumeSpecName: "util") pod "f107fa8c-f81f-4d3f-8a14-e718942491b8" (UID: "f107fa8c-f81f-4d3f-8a14-e718942491b8"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:29:49 crc kubenswrapper[4733]: I0202 15:29:49.008638 4733 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f107fa8c-f81f-4d3f-8a14-e718942491b8-util\") on node \"crc\" DevicePath \"\"" Feb 02 15:29:49 crc kubenswrapper[4733]: I0202 15:29:49.008700 4733 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f107fa8c-f81f-4d3f-8a14-e718942491b8-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 15:29:49 crc kubenswrapper[4733]: I0202 15:29:49.008725 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wltk7\" (UniqueName: \"kubernetes.io/projected/f107fa8c-f81f-4d3f-8a14-e718942491b8-kube-api-access-wltk7\") on node \"crc\" DevicePath \"\"" Feb 02 15:29:49 crc kubenswrapper[4733]: I0202 15:29:49.539542 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f75764trqm" event={"ID":"f107fa8c-f81f-4d3f-8a14-e718942491b8","Type":"ContainerDied","Data":"d945384e860d5e02676c1fe38f2f3e835220dbceda859d2fd9295268a7255d09"} Feb 02 15:29:49 crc kubenswrapper[4733]: I0202 15:29:49.539597 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d945384e860d5e02676c1fe38f2f3e835220dbceda859d2fd9295268a7255d09" Feb 02 15:29:49 crc kubenswrapper[4733]: I0202 15:29:49.539641 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f75764trqm" Feb 02 15:29:59 crc kubenswrapper[4733]: I0202 15:29:59.474596 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-b9d5f6fc5-nrtgx"] Feb 02 15:29:59 crc kubenswrapper[4733]: E0202 15:29:59.475383 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f107fa8c-f81f-4d3f-8a14-e718942491b8" containerName="util" Feb 02 15:29:59 crc kubenswrapper[4733]: I0202 15:29:59.475401 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f107fa8c-f81f-4d3f-8a14-e718942491b8" containerName="util" Feb 02 15:29:59 crc kubenswrapper[4733]: E0202 15:29:59.475416 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f107fa8c-f81f-4d3f-8a14-e718942491b8" containerName="pull" Feb 02 15:29:59 crc kubenswrapper[4733]: I0202 15:29:59.475424 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f107fa8c-f81f-4d3f-8a14-e718942491b8" containerName="pull" Feb 02 15:29:59 crc kubenswrapper[4733]: E0202 15:29:59.475439 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f107fa8c-f81f-4d3f-8a14-e718942491b8" containerName="extract" Feb 02 15:29:59 crc kubenswrapper[4733]: I0202 15:29:59.475448 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f107fa8c-f81f-4d3f-8a14-e718942491b8" containerName="extract" Feb 02 15:29:59 crc kubenswrapper[4733]: I0202 15:29:59.475569 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f107fa8c-f81f-4d3f-8a14-e718942491b8" containerName="extract" Feb 02 15:29:59 crc kubenswrapper[4733]: I0202 15:29:59.476016 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-b9d5f6fc5-nrtgx" Feb 02 15:29:59 crc kubenswrapper[4733]: I0202 15:29:59.482256 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-service-cert" Feb 02 15:29:59 crc kubenswrapper[4733]: I0202 15:29:59.488823 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-fs5hs" Feb 02 15:29:59 crc kubenswrapper[4733]: I0202 15:29:59.506903 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-b9d5f6fc5-nrtgx"] Feb 02 15:29:59 crc kubenswrapper[4733]: I0202 15:29:59.653690 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqpqq\" (UniqueName: \"kubernetes.io/projected/7d80c76f-a56e-4c8b-9a02-73311129ea44-kube-api-access-nqpqq\") pod \"infra-operator-controller-manager-b9d5f6fc5-nrtgx\" (UID: \"7d80c76f-a56e-4c8b-9a02-73311129ea44\") " pod="openstack-operators/infra-operator-controller-manager-b9d5f6fc5-nrtgx" Feb 02 15:29:59 crc kubenswrapper[4733]: I0202 15:29:59.653747 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7d80c76f-a56e-4c8b-9a02-73311129ea44-webhook-cert\") pod \"infra-operator-controller-manager-b9d5f6fc5-nrtgx\" (UID: \"7d80c76f-a56e-4c8b-9a02-73311129ea44\") " pod="openstack-operators/infra-operator-controller-manager-b9d5f6fc5-nrtgx" Feb 02 15:29:59 crc kubenswrapper[4733]: I0202 15:29:59.654187 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7d80c76f-a56e-4c8b-9a02-73311129ea44-apiservice-cert\") pod \"infra-operator-controller-manager-b9d5f6fc5-nrtgx\" (UID: \"7d80c76f-a56e-4c8b-9a02-73311129ea44\") " pod="openstack-operators/infra-operator-controller-manager-b9d5f6fc5-nrtgx" Feb 02 15:29:59 crc kubenswrapper[4733]: I0202 15:29:59.756013 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7d80c76f-a56e-4c8b-9a02-73311129ea44-apiservice-cert\") pod \"infra-operator-controller-manager-b9d5f6fc5-nrtgx\" (UID: \"7d80c76f-a56e-4c8b-9a02-73311129ea44\") " pod="openstack-operators/infra-operator-controller-manager-b9d5f6fc5-nrtgx" Feb 02 15:29:59 crc kubenswrapper[4733]: I0202 15:29:59.756115 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqpqq\" (UniqueName: \"kubernetes.io/projected/7d80c76f-a56e-4c8b-9a02-73311129ea44-kube-api-access-nqpqq\") pod \"infra-operator-controller-manager-b9d5f6fc5-nrtgx\" (UID: \"7d80c76f-a56e-4c8b-9a02-73311129ea44\") " pod="openstack-operators/infra-operator-controller-manager-b9d5f6fc5-nrtgx" Feb 02 15:29:59 crc kubenswrapper[4733]: I0202 15:29:59.756148 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7d80c76f-a56e-4c8b-9a02-73311129ea44-webhook-cert\") pod \"infra-operator-controller-manager-b9d5f6fc5-nrtgx\" (UID: \"7d80c76f-a56e-4c8b-9a02-73311129ea44\") " pod="openstack-operators/infra-operator-controller-manager-b9d5f6fc5-nrtgx" Feb 02 15:29:59 crc kubenswrapper[4733]: I0202 15:29:59.764177 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7d80c76f-a56e-4c8b-9a02-73311129ea44-apiservice-cert\") pod \"infra-operator-controller-manager-b9d5f6fc5-nrtgx\" (UID: \"7d80c76f-a56e-4c8b-9a02-73311129ea44\") " pod="openstack-operators/infra-operator-controller-manager-b9d5f6fc5-nrtgx" Feb 02 15:29:59 crc kubenswrapper[4733]: I0202 15:29:59.764584 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7d80c76f-a56e-4c8b-9a02-73311129ea44-webhook-cert\") pod \"infra-operator-controller-manager-b9d5f6fc5-nrtgx\" (UID: \"7d80c76f-a56e-4c8b-9a02-73311129ea44\") " pod="openstack-operators/infra-operator-controller-manager-b9d5f6fc5-nrtgx" Feb 02 15:29:59 crc kubenswrapper[4733]: I0202 15:29:59.787499 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqpqq\" (UniqueName: \"kubernetes.io/projected/7d80c76f-a56e-4c8b-9a02-73311129ea44-kube-api-access-nqpqq\") pod \"infra-operator-controller-manager-b9d5f6fc5-nrtgx\" (UID: \"7d80c76f-a56e-4c8b-9a02-73311129ea44\") " pod="openstack-operators/infra-operator-controller-manager-b9d5f6fc5-nrtgx" Feb 02 15:29:59 crc kubenswrapper[4733]: I0202 15:29:59.795546 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-b9d5f6fc5-nrtgx" Feb 02 15:30:00 crc kubenswrapper[4733]: I0202 15:30:00.156544 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500770-r6bdp"] Feb 02 15:30:00 crc kubenswrapper[4733]: I0202 15:30:00.157762 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500770-r6bdp" Feb 02 15:30:00 crc kubenswrapper[4733]: I0202 15:30:00.160295 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 02 15:30:00 crc kubenswrapper[4733]: I0202 15:30:00.160632 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 02 15:30:00 crc kubenswrapper[4733]: I0202 15:30:00.168119 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500770-r6bdp"] Feb 02 15:30:00 crc kubenswrapper[4733]: I0202 15:30:00.218442 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-b9d5f6fc5-nrtgx"] Feb 02 15:30:00 crc kubenswrapper[4733]: W0202 15:30:00.224616 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7d80c76f_a56e_4c8b_9a02_73311129ea44.slice/crio-22507597bb0f5d49c018769e0899bc2cc2c58f459a851928b8c408cfd6ca98aa WatchSource:0}: Error finding container 22507597bb0f5d49c018769e0899bc2cc2c58f459a851928b8c408cfd6ca98aa: Status 404 returned error can't find the container with id 22507597bb0f5d49c018769e0899bc2cc2c58f459a851928b8c408cfd6ca98aa Feb 02 15:30:00 crc kubenswrapper[4733]: I0202 15:30:00.264002 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/28575f75-56bc-433b-80a7-3ca1030ca8d0-config-volume\") pod \"collect-profiles-29500770-r6bdp\" (UID: \"28575f75-56bc-433b-80a7-3ca1030ca8d0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500770-r6bdp" Feb 02 15:30:00 crc kubenswrapper[4733]: I0202 15:30:00.264059 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/28575f75-56bc-433b-80a7-3ca1030ca8d0-secret-volume\") pod \"collect-profiles-29500770-r6bdp\" (UID: \"28575f75-56bc-433b-80a7-3ca1030ca8d0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500770-r6bdp" Feb 02 15:30:00 crc kubenswrapper[4733]: I0202 15:30:00.264129 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtpqd\" (UniqueName: \"kubernetes.io/projected/28575f75-56bc-433b-80a7-3ca1030ca8d0-kube-api-access-wtpqd\") pod \"collect-profiles-29500770-r6bdp\" (UID: \"28575f75-56bc-433b-80a7-3ca1030ca8d0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500770-r6bdp" Feb 02 15:30:00 crc kubenswrapper[4733]: I0202 15:30:00.365623 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtpqd\" (UniqueName: \"kubernetes.io/projected/28575f75-56bc-433b-80a7-3ca1030ca8d0-kube-api-access-wtpqd\") pod \"collect-profiles-29500770-r6bdp\" (UID: \"28575f75-56bc-433b-80a7-3ca1030ca8d0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500770-r6bdp" Feb 02 15:30:00 crc kubenswrapper[4733]: I0202 15:30:00.365910 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/28575f75-56bc-433b-80a7-3ca1030ca8d0-config-volume\") pod \"collect-profiles-29500770-r6bdp\" (UID: \"28575f75-56bc-433b-80a7-3ca1030ca8d0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500770-r6bdp" Feb 02 15:30:00 crc kubenswrapper[4733]: I0202 15:30:00.365981 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/28575f75-56bc-433b-80a7-3ca1030ca8d0-secret-volume\") pod \"collect-profiles-29500770-r6bdp\" (UID: \"28575f75-56bc-433b-80a7-3ca1030ca8d0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500770-r6bdp" Feb 02 15:30:00 crc kubenswrapper[4733]: I0202 15:30:00.366939 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/28575f75-56bc-433b-80a7-3ca1030ca8d0-config-volume\") pod \"collect-profiles-29500770-r6bdp\" (UID: \"28575f75-56bc-433b-80a7-3ca1030ca8d0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500770-r6bdp" Feb 02 15:30:00 crc kubenswrapper[4733]: I0202 15:30:00.372503 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/28575f75-56bc-433b-80a7-3ca1030ca8d0-secret-volume\") pod \"collect-profiles-29500770-r6bdp\" (UID: \"28575f75-56bc-433b-80a7-3ca1030ca8d0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500770-r6bdp" Feb 02 15:30:00 crc kubenswrapper[4733]: I0202 15:30:00.382599 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtpqd\" (UniqueName: \"kubernetes.io/projected/28575f75-56bc-433b-80a7-3ca1030ca8d0-kube-api-access-wtpqd\") pod \"collect-profiles-29500770-r6bdp\" (UID: \"28575f75-56bc-433b-80a7-3ca1030ca8d0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500770-r6bdp" Feb 02 15:30:00 crc kubenswrapper[4733]: I0202 15:30:00.479359 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500770-r6bdp" Feb 02 15:30:00 crc kubenswrapper[4733]: I0202 15:30:00.622499 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-b9d5f6fc5-nrtgx" event={"ID":"7d80c76f-a56e-4c8b-9a02-73311129ea44","Type":"ContainerStarted","Data":"22507597bb0f5d49c018769e0899bc2cc2c58f459a851928b8c408cfd6ca98aa"} Feb 02 15:30:00 crc kubenswrapper[4733]: I0202 15:30:00.890888 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500770-r6bdp"] Feb 02 15:30:01 crc kubenswrapper[4733]: I0202 15:30:01.635776 4733 generic.go:334] "Generic (PLEG): container finished" podID="28575f75-56bc-433b-80a7-3ca1030ca8d0" containerID="0a3fcc1305489ab13625543763254bfbf25e4f424312c281381d377629053eb8" exitCode=0 Feb 02 15:30:01 crc kubenswrapper[4733]: I0202 15:30:01.635818 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500770-r6bdp" event={"ID":"28575f75-56bc-433b-80a7-3ca1030ca8d0","Type":"ContainerDied","Data":"0a3fcc1305489ab13625543763254bfbf25e4f424312c281381d377629053eb8"} Feb 02 15:30:01 crc kubenswrapper[4733]: I0202 15:30:01.636007 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500770-r6bdp" event={"ID":"28575f75-56bc-433b-80a7-3ca1030ca8d0","Type":"ContainerStarted","Data":"55849a224329b0c1ce214eb58b9c5ef251796a14e48295515a0d31a188f0ed4c"} Feb 02 15:30:02 crc kubenswrapper[4733]: I0202 15:30:02.645035 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-b9d5f6fc5-nrtgx" event={"ID":"7d80c76f-a56e-4c8b-9a02-73311129ea44","Type":"ContainerStarted","Data":"59e0e2d29d977f2cc61350355cd4e39bef9cd65be1e1ab4b2dfd1c4e70dbfa7a"} Feb 02 15:30:02 crc kubenswrapper[4733]: I0202 15:30:02.665909 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-b9d5f6fc5-nrtgx" podStartSLOduration=1.763350053 podStartE2EDuration="3.665888733s" podCreationTimestamp="2026-02-02 15:29:59 +0000 UTC" firstStartedPulling="2026-02-02 15:30:00.227228008 +0000 UTC m=+883.678689366" lastFinishedPulling="2026-02-02 15:30:02.129766688 +0000 UTC m=+885.581228046" observedRunningTime="2026-02-02 15:30:02.663608794 +0000 UTC m=+886.115070152" watchObservedRunningTime="2026-02-02 15:30:02.665888733 +0000 UTC m=+886.117350101" Feb 02 15:30:02 crc kubenswrapper[4733]: I0202 15:30:02.936214 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500770-r6bdp" Feb 02 15:30:03 crc kubenswrapper[4733]: I0202 15:30:03.121008 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/28575f75-56bc-433b-80a7-3ca1030ca8d0-config-volume\") pod \"28575f75-56bc-433b-80a7-3ca1030ca8d0\" (UID: \"28575f75-56bc-433b-80a7-3ca1030ca8d0\") " Feb 02 15:30:03 crc kubenswrapper[4733]: I0202 15:30:03.121127 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wtpqd\" (UniqueName: \"kubernetes.io/projected/28575f75-56bc-433b-80a7-3ca1030ca8d0-kube-api-access-wtpqd\") pod \"28575f75-56bc-433b-80a7-3ca1030ca8d0\" (UID: \"28575f75-56bc-433b-80a7-3ca1030ca8d0\") " Feb 02 15:30:03 crc kubenswrapper[4733]: I0202 15:30:03.121185 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/28575f75-56bc-433b-80a7-3ca1030ca8d0-secret-volume\") pod \"28575f75-56bc-433b-80a7-3ca1030ca8d0\" (UID: \"28575f75-56bc-433b-80a7-3ca1030ca8d0\") " Feb 02 15:30:03 crc kubenswrapper[4733]: I0202 15:30:03.121942 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28575f75-56bc-433b-80a7-3ca1030ca8d0-config-volume" (OuterVolumeSpecName: "config-volume") pod "28575f75-56bc-433b-80a7-3ca1030ca8d0" (UID: "28575f75-56bc-433b-80a7-3ca1030ca8d0"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:30:03 crc kubenswrapper[4733]: I0202 15:30:03.126909 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28575f75-56bc-433b-80a7-3ca1030ca8d0-kube-api-access-wtpqd" (OuterVolumeSpecName: "kube-api-access-wtpqd") pod "28575f75-56bc-433b-80a7-3ca1030ca8d0" (UID: "28575f75-56bc-433b-80a7-3ca1030ca8d0"). InnerVolumeSpecName "kube-api-access-wtpqd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:30:03 crc kubenswrapper[4733]: I0202 15:30:03.127535 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28575f75-56bc-433b-80a7-3ca1030ca8d0-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "28575f75-56bc-433b-80a7-3ca1030ca8d0" (UID: "28575f75-56bc-433b-80a7-3ca1030ca8d0"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:30:03 crc kubenswrapper[4733]: I0202 15:30:03.222988 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wtpqd\" (UniqueName: \"kubernetes.io/projected/28575f75-56bc-433b-80a7-3ca1030ca8d0-kube-api-access-wtpqd\") on node \"crc\" DevicePath \"\"" Feb 02 15:30:03 crc kubenswrapper[4733]: I0202 15:30:03.223021 4733 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/28575f75-56bc-433b-80a7-3ca1030ca8d0-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 02 15:30:03 crc kubenswrapper[4733]: I0202 15:30:03.223032 4733 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/28575f75-56bc-433b-80a7-3ca1030ca8d0-config-volume\") on node \"crc\" DevicePath \"\"" Feb 02 15:30:03 crc kubenswrapper[4733]: I0202 15:30:03.651256 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500770-r6bdp" Feb 02 15:30:03 crc kubenswrapper[4733]: I0202 15:30:03.651246 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500770-r6bdp" event={"ID":"28575f75-56bc-433b-80a7-3ca1030ca8d0","Type":"ContainerDied","Data":"55849a224329b0c1ce214eb58b9c5ef251796a14e48295515a0d31a188f0ed4c"} Feb 02 15:30:03 crc kubenswrapper[4733]: I0202 15:30:03.651730 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="55849a224329b0c1ce214eb58b9c5ef251796a14e48295515a0d31a188f0ed4c" Feb 02 15:30:03 crc kubenswrapper[4733]: I0202 15:30:03.651785 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-b9d5f6fc5-nrtgx" Feb 02 15:30:05 crc kubenswrapper[4733]: I0202 15:30:05.794814 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["manila-kuttl-tests/openstack-galera-0"] Feb 02 15:30:05 crc kubenswrapper[4733]: E0202 15:30:05.795107 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28575f75-56bc-433b-80a7-3ca1030ca8d0" containerName="collect-profiles" Feb 02 15:30:05 crc kubenswrapper[4733]: I0202 15:30:05.795123 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="28575f75-56bc-433b-80a7-3ca1030ca8d0" containerName="collect-profiles" Feb 02 15:30:05 crc kubenswrapper[4733]: I0202 15:30:05.795280 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="28575f75-56bc-433b-80a7-3ca1030ca8d0" containerName="collect-profiles" Feb 02 15:30:05 crc kubenswrapper[4733]: I0202 15:30:05.795939 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/openstack-galera-0" Feb 02 15:30:05 crc kubenswrapper[4733]: I0202 15:30:05.798951 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"manila-kuttl-tests"/"openshift-service-ca.crt" Feb 02 15:30:05 crc kubenswrapper[4733]: I0202 15:30:05.799030 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"manila-kuttl-tests"/"openstack-config-data" Feb 02 15:30:05 crc kubenswrapper[4733]: I0202 15:30:05.799680 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"manila-kuttl-tests"/"kube-root-ca.crt" Feb 02 15:30:05 crc kubenswrapper[4733]: I0202 15:30:05.800081 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"galera-openstack-dockercfg-4hmlb" Feb 02 15:30:05 crc kubenswrapper[4733]: I0202 15:30:05.802508 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"manila-kuttl-tests"/"openstack-scripts" Feb 02 15:30:05 crc kubenswrapper[4733]: I0202 15:30:05.806761 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["manila-kuttl-tests/openstack-galera-2"] Feb 02 15:30:05 crc kubenswrapper[4733]: I0202 15:30:05.807924 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/openstack-galera-2" Feb 02 15:30:05 crc kubenswrapper[4733]: I0202 15:30:05.812921 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["manila-kuttl-tests/openstack-galera-1"] Feb 02 15:30:05 crc kubenswrapper[4733]: I0202 15:30:05.814011 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/openstack-galera-1" Feb 02 15:30:05 crc kubenswrapper[4733]: I0202 15:30:05.822509 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/openstack-galera-0"] Feb 02 15:30:05 crc kubenswrapper[4733]: I0202 15:30:05.852513 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/openstack-galera-2"] Feb 02 15:30:05 crc kubenswrapper[4733]: I0202 15:30:05.872991 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/openstack-galera-1"] Feb 02 15:30:05 crc kubenswrapper[4733]: I0202 15:30:05.957106 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/02261047-2e53-4b34-bde5-5da7facb243c-config-data-default\") pod \"openstack-galera-2\" (UID: \"02261047-2e53-4b34-bde5-5da7facb243c\") " pod="manila-kuttl-tests/openstack-galera-2" Feb 02 15:30:05 crc kubenswrapper[4733]: I0202 15:30:05.957196 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f177ebf9-728b-4ba2-b9b7-6c676f9d014a-operator-scripts\") pod \"openstack-galera-0\" (UID: \"f177ebf9-728b-4ba2-b9b7-6c676f9d014a\") " pod="manila-kuttl-tests/openstack-galera-0" Feb 02 15:30:05 crc kubenswrapper[4733]: I0202 15:30:05.957233 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d63f3e93-f45f-4f7d-a29b-2eaae536fa5d-config-data-generated\") pod \"openstack-galera-1\" (UID: \"d63f3e93-f45f-4f7d-a29b-2eaae536fa5d\") " pod="manila-kuttl-tests/openstack-galera-1" Feb 02 15:30:05 crc kubenswrapper[4733]: I0202 15:30:05.957267 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f177ebf9-728b-4ba2-b9b7-6c676f9d014a-config-data-generated\") pod \"openstack-galera-0\" (UID: \"f177ebf9-728b-4ba2-b9b7-6c676f9d014a\") " pod="manila-kuttl-tests/openstack-galera-0" Feb 02 15:30:05 crc kubenswrapper[4733]: I0202 15:30:05.957368 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"f177ebf9-728b-4ba2-b9b7-6c676f9d014a\") " pod="manila-kuttl-tests/openstack-galera-0" Feb 02 15:30:05 crc kubenswrapper[4733]: I0202 15:30:05.957451 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/02261047-2e53-4b34-bde5-5da7facb243c-kolla-config\") pod \"openstack-galera-2\" (UID: \"02261047-2e53-4b34-bde5-5da7facb243c\") " pod="manila-kuttl-tests/openstack-galera-2" Feb 02 15:30:05 crc kubenswrapper[4733]: I0202 15:30:05.957490 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f177ebf9-728b-4ba2-b9b7-6c676f9d014a-kolla-config\") pod \"openstack-galera-0\" (UID: \"f177ebf9-728b-4ba2-b9b7-6c676f9d014a\") " pod="manila-kuttl-tests/openstack-galera-0" Feb 02 15:30:05 crc kubenswrapper[4733]: I0202 15:30:05.957523 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4kxk\" (UniqueName: \"kubernetes.io/projected/f177ebf9-728b-4ba2-b9b7-6c676f9d014a-kube-api-access-f4kxk\") pod \"openstack-galera-0\" (UID: \"f177ebf9-728b-4ba2-b9b7-6c676f9d014a\") " pod="manila-kuttl-tests/openstack-galera-0" Feb 02 15:30:05 crc kubenswrapper[4733]: I0202 15:30:05.957552 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d63f3e93-f45f-4f7d-a29b-2eaae536fa5d-operator-scripts\") pod \"openstack-galera-1\" (UID: \"d63f3e93-f45f-4f7d-a29b-2eaae536fa5d\") " pod="manila-kuttl-tests/openstack-galera-1" Feb 02 15:30:05 crc kubenswrapper[4733]: I0202 15:30:05.957653 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6q42z\" (UniqueName: \"kubernetes.io/projected/02261047-2e53-4b34-bde5-5da7facb243c-kube-api-access-6q42z\") pod \"openstack-galera-2\" (UID: \"02261047-2e53-4b34-bde5-5da7facb243c\") " pod="manila-kuttl-tests/openstack-galera-2" Feb 02 15:30:05 crc kubenswrapper[4733]: I0202 15:30:05.957764 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f177ebf9-728b-4ba2-b9b7-6c676f9d014a-config-data-default\") pod \"openstack-galera-0\" (UID: \"f177ebf9-728b-4ba2-b9b7-6c676f9d014a\") " pod="manila-kuttl-tests/openstack-galera-0" Feb 02 15:30:05 crc kubenswrapper[4733]: I0202 15:30:05.957819 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/02261047-2e53-4b34-bde5-5da7facb243c-operator-scripts\") pod \"openstack-galera-2\" (UID: \"02261047-2e53-4b34-bde5-5da7facb243c\") " pod="manila-kuttl-tests/openstack-galera-2" Feb 02 15:30:05 crc kubenswrapper[4733]: I0202 15:30:05.957868 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d63f3e93-f45f-4f7d-a29b-2eaae536fa5d-kolla-config\") pod \"openstack-galera-1\" (UID: \"d63f3e93-f45f-4f7d-a29b-2eaae536fa5d\") " pod="manila-kuttl-tests/openstack-galera-1" Feb 02 15:30:05 crc kubenswrapper[4733]: I0202 15:30:05.957895 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-2\" (UID: \"02261047-2e53-4b34-bde5-5da7facb243c\") " pod="manila-kuttl-tests/openstack-galera-2" Feb 02 15:30:05 crc kubenswrapper[4733]: I0202 15:30:05.957932 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xsq68\" (UniqueName: \"kubernetes.io/projected/d63f3e93-f45f-4f7d-a29b-2eaae536fa5d-kube-api-access-xsq68\") pod \"openstack-galera-1\" (UID: \"d63f3e93-f45f-4f7d-a29b-2eaae536fa5d\") " pod="manila-kuttl-tests/openstack-galera-1" Feb 02 15:30:05 crc kubenswrapper[4733]: I0202 15:30:05.958028 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d63f3e93-f45f-4f7d-a29b-2eaae536fa5d-config-data-default\") pod \"openstack-galera-1\" (UID: \"d63f3e93-f45f-4f7d-a29b-2eaae536fa5d\") " pod="manila-kuttl-tests/openstack-galera-1" Feb 02 15:30:05 crc kubenswrapper[4733]: I0202 15:30:05.958108 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/02261047-2e53-4b34-bde5-5da7facb243c-config-data-generated\") pod \"openstack-galera-2\" (UID: \"02261047-2e53-4b34-bde5-5da7facb243c\") " pod="manila-kuttl-tests/openstack-galera-2" Feb 02 15:30:05 crc kubenswrapper[4733]: I0202 15:30:05.958134 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-1\" (UID: \"d63f3e93-f45f-4f7d-a29b-2eaae536fa5d\") " pod="manila-kuttl-tests/openstack-galera-1" Feb 02 15:30:06 crc kubenswrapper[4733]: I0202 15:30:06.059298 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f177ebf9-728b-4ba2-b9b7-6c676f9d014a-config-data-default\") pod \"openstack-galera-0\" (UID: \"f177ebf9-728b-4ba2-b9b7-6c676f9d014a\") " pod="manila-kuttl-tests/openstack-galera-0" Feb 02 15:30:06 crc kubenswrapper[4733]: I0202 15:30:06.059644 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/02261047-2e53-4b34-bde5-5da7facb243c-operator-scripts\") pod \"openstack-galera-2\" (UID: \"02261047-2e53-4b34-bde5-5da7facb243c\") " pod="manila-kuttl-tests/openstack-galera-2" Feb 02 15:30:06 crc kubenswrapper[4733]: I0202 15:30:06.059678 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d63f3e93-f45f-4f7d-a29b-2eaae536fa5d-kolla-config\") pod \"openstack-galera-1\" (UID: \"d63f3e93-f45f-4f7d-a29b-2eaae536fa5d\") " pod="manila-kuttl-tests/openstack-galera-1" Feb 02 15:30:06 crc kubenswrapper[4733]: I0202 15:30:06.059703 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-2\" (UID: \"02261047-2e53-4b34-bde5-5da7facb243c\") " pod="manila-kuttl-tests/openstack-galera-2" Feb 02 15:30:06 crc kubenswrapper[4733]: I0202 15:30:06.059732 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xsq68\" (UniqueName: \"kubernetes.io/projected/d63f3e93-f45f-4f7d-a29b-2eaae536fa5d-kube-api-access-xsq68\") pod \"openstack-galera-1\" (UID: \"d63f3e93-f45f-4f7d-a29b-2eaae536fa5d\") " pod="manila-kuttl-tests/openstack-galera-1" Feb 02 15:30:06 crc kubenswrapper[4733]: I0202 15:30:06.059754 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d63f3e93-f45f-4f7d-a29b-2eaae536fa5d-config-data-default\") pod \"openstack-galera-1\" (UID: \"d63f3e93-f45f-4f7d-a29b-2eaae536fa5d\") " pod="manila-kuttl-tests/openstack-galera-1" Feb 02 15:30:06 crc kubenswrapper[4733]: I0202 15:30:06.059786 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/02261047-2e53-4b34-bde5-5da7facb243c-config-data-generated\") pod \"openstack-galera-2\" (UID: \"02261047-2e53-4b34-bde5-5da7facb243c\") " pod="manila-kuttl-tests/openstack-galera-2" Feb 02 15:30:06 crc kubenswrapper[4733]: I0202 15:30:06.059807 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-1\" (UID: \"d63f3e93-f45f-4f7d-a29b-2eaae536fa5d\") " pod="manila-kuttl-tests/openstack-galera-1" Feb 02 15:30:06 crc kubenswrapper[4733]: I0202 15:30:06.059866 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/02261047-2e53-4b34-bde5-5da7facb243c-config-data-default\") pod \"openstack-galera-2\" (UID: \"02261047-2e53-4b34-bde5-5da7facb243c\") " pod="manila-kuttl-tests/openstack-galera-2" Feb 02 15:30:06 crc kubenswrapper[4733]: I0202 15:30:06.059894 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f177ebf9-728b-4ba2-b9b7-6c676f9d014a-operator-scripts\") pod \"openstack-galera-0\" (UID: \"f177ebf9-728b-4ba2-b9b7-6c676f9d014a\") " pod="manila-kuttl-tests/openstack-galera-0" Feb 02 15:30:06 crc kubenswrapper[4733]: I0202 15:30:06.059924 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d63f3e93-f45f-4f7d-a29b-2eaae536fa5d-config-data-generated\") pod \"openstack-galera-1\" (UID: \"d63f3e93-f45f-4f7d-a29b-2eaae536fa5d\") " pod="manila-kuttl-tests/openstack-galera-1" Feb 02 15:30:06 crc kubenswrapper[4733]: I0202 15:30:06.059964 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f177ebf9-728b-4ba2-b9b7-6c676f9d014a-config-data-generated\") pod \"openstack-galera-0\" (UID: \"f177ebf9-728b-4ba2-b9b7-6c676f9d014a\") " pod="manila-kuttl-tests/openstack-galera-0" Feb 02 15:30:06 crc kubenswrapper[4733]: I0202 15:30:06.059997 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"f177ebf9-728b-4ba2-b9b7-6c676f9d014a\") " pod="manila-kuttl-tests/openstack-galera-0" Feb 02 15:30:06 crc kubenswrapper[4733]: I0202 15:30:06.060024 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/02261047-2e53-4b34-bde5-5da7facb243c-kolla-config\") pod \"openstack-galera-2\" (UID: \"02261047-2e53-4b34-bde5-5da7facb243c\") " pod="manila-kuttl-tests/openstack-galera-2" Feb 02 15:30:06 crc kubenswrapper[4733]: I0202 15:30:06.060054 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f177ebf9-728b-4ba2-b9b7-6c676f9d014a-kolla-config\") pod \"openstack-galera-0\" (UID: \"f177ebf9-728b-4ba2-b9b7-6c676f9d014a\") " pod="manila-kuttl-tests/openstack-galera-0" Feb 02 15:30:06 crc kubenswrapper[4733]: I0202 15:30:06.060085 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4kxk\" (UniqueName: \"kubernetes.io/projected/f177ebf9-728b-4ba2-b9b7-6c676f9d014a-kube-api-access-f4kxk\") pod \"openstack-galera-0\" (UID: \"f177ebf9-728b-4ba2-b9b7-6c676f9d014a\") " pod="manila-kuttl-tests/openstack-galera-0" Feb 02 15:30:06 crc kubenswrapper[4733]: I0202 15:30:06.060113 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d63f3e93-f45f-4f7d-a29b-2eaae536fa5d-operator-scripts\") pod \"openstack-galera-1\" (UID: \"d63f3e93-f45f-4f7d-a29b-2eaae536fa5d\") " pod="manila-kuttl-tests/openstack-galera-1" Feb 02 15:30:06 crc kubenswrapper[4733]: I0202 15:30:06.060141 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f177ebf9-728b-4ba2-b9b7-6c676f9d014a-config-data-default\") pod \"openstack-galera-0\" (UID: \"f177ebf9-728b-4ba2-b9b7-6c676f9d014a\") " pod="manila-kuttl-tests/openstack-galera-0" Feb 02 15:30:06 crc kubenswrapper[4733]: I0202 15:30:06.060147 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6q42z\" (UniqueName: \"kubernetes.io/projected/02261047-2e53-4b34-bde5-5da7facb243c-kube-api-access-6q42z\") pod \"openstack-galera-2\" (UID: \"02261047-2e53-4b34-bde5-5da7facb243c\") " pod="manila-kuttl-tests/openstack-galera-2" Feb 02 15:30:06 crc kubenswrapper[4733]: I0202 15:30:06.060340 4733 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-2\" (UID: \"02261047-2e53-4b34-bde5-5da7facb243c\") device mount path \"/mnt/openstack/pv10\"" pod="manila-kuttl-tests/openstack-galera-2" Feb 02 15:30:06 crc kubenswrapper[4733]: I0202 15:30:06.060451 4733 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"f177ebf9-728b-4ba2-b9b7-6c676f9d014a\") device mount path \"/mnt/openstack/pv04\"" pod="manila-kuttl-tests/openstack-galera-0" Feb 02 15:30:06 crc kubenswrapper[4733]: I0202 15:30:06.060605 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/02261047-2e53-4b34-bde5-5da7facb243c-config-data-generated\") pod \"openstack-galera-2\" (UID: \"02261047-2e53-4b34-bde5-5da7facb243c\") " pod="manila-kuttl-tests/openstack-galera-2" Feb 02 15:30:06 crc kubenswrapper[4733]: I0202 15:30:06.060897 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d63f3e93-f45f-4f7d-a29b-2eaae536fa5d-kolla-config\") pod \"openstack-galera-1\" (UID: \"d63f3e93-f45f-4f7d-a29b-2eaae536fa5d\") " pod="manila-kuttl-tests/openstack-galera-1" Feb 02 15:30:06 crc kubenswrapper[4733]: I0202 15:30:06.061357 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/02261047-2e53-4b34-bde5-5da7facb243c-kolla-config\") pod \"openstack-galera-2\" (UID: \"02261047-2e53-4b34-bde5-5da7facb243c\") " pod="manila-kuttl-tests/openstack-galera-2" Feb 02 15:30:06 crc kubenswrapper[4733]: I0202 15:30:06.061395 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/02261047-2e53-4b34-bde5-5da7facb243c-operator-scripts\") pod \"openstack-galera-2\" (UID: \"02261047-2e53-4b34-bde5-5da7facb243c\") " pod="manila-kuttl-tests/openstack-galera-2" Feb 02 15:30:06 crc kubenswrapper[4733]: I0202 15:30:06.061423 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d63f3e93-f45f-4f7d-a29b-2eaae536fa5d-config-data-default\") pod \"openstack-galera-1\" (UID: \"d63f3e93-f45f-4f7d-a29b-2eaae536fa5d\") " pod="manila-kuttl-tests/openstack-galera-1" Feb 02 15:30:06 crc kubenswrapper[4733]: I0202 15:30:06.061646 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f177ebf9-728b-4ba2-b9b7-6c676f9d014a-kolla-config\") pod \"openstack-galera-0\" (UID: \"f177ebf9-728b-4ba2-b9b7-6c676f9d014a\") " pod="manila-kuttl-tests/openstack-galera-0" Feb 02 15:30:06 crc kubenswrapper[4733]: I0202 15:30:06.061747 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d63f3e93-f45f-4f7d-a29b-2eaae536fa5d-config-data-generated\") pod \"openstack-galera-1\" (UID: \"d63f3e93-f45f-4f7d-a29b-2eaae536fa5d\") " pod="manila-kuttl-tests/openstack-galera-1" Feb 02 15:30:06 crc kubenswrapper[4733]: I0202 15:30:06.061818 4733 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-1\" (UID: \"d63f3e93-f45f-4f7d-a29b-2eaae536fa5d\") device mount path \"/mnt/openstack/pv02\"" pod="manila-kuttl-tests/openstack-galera-1" Feb 02 15:30:06 crc kubenswrapper[4733]: I0202 15:30:06.061875 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f177ebf9-728b-4ba2-b9b7-6c676f9d014a-config-data-generated\") pod \"openstack-galera-0\" (UID: \"f177ebf9-728b-4ba2-b9b7-6c676f9d014a\") " pod="manila-kuttl-tests/openstack-galera-0" Feb 02 15:30:06 crc kubenswrapper[4733]: I0202 15:30:06.062549 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f177ebf9-728b-4ba2-b9b7-6c676f9d014a-operator-scripts\") pod \"openstack-galera-0\" (UID: \"f177ebf9-728b-4ba2-b9b7-6c676f9d014a\") " pod="manila-kuttl-tests/openstack-galera-0" Feb 02 15:30:06 crc kubenswrapper[4733]: I0202 15:30:06.062928 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/02261047-2e53-4b34-bde5-5da7facb243c-config-data-default\") pod \"openstack-galera-2\" (UID: \"02261047-2e53-4b34-bde5-5da7facb243c\") " pod="manila-kuttl-tests/openstack-galera-2" Feb 02 15:30:06 crc kubenswrapper[4733]: I0202 15:30:06.063384 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d63f3e93-f45f-4f7d-a29b-2eaae536fa5d-operator-scripts\") pod \"openstack-galera-1\" (UID: \"d63f3e93-f45f-4f7d-a29b-2eaae536fa5d\") " pod="manila-kuttl-tests/openstack-galera-1" Feb 02 15:30:06 crc kubenswrapper[4733]: I0202 15:30:06.092033 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-2\" (UID: \"02261047-2e53-4b34-bde5-5da7facb243c\") " pod="manila-kuttl-tests/openstack-galera-2" Feb 02 15:30:06 crc kubenswrapper[4733]: I0202 15:30:06.094133 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"f177ebf9-728b-4ba2-b9b7-6c676f9d014a\") " pod="manila-kuttl-tests/openstack-galera-0" Feb 02 15:30:06 crc kubenswrapper[4733]: I0202 15:30:06.097759 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-1\" (UID: \"d63f3e93-f45f-4f7d-a29b-2eaae536fa5d\") " pod="manila-kuttl-tests/openstack-galera-1" Feb 02 15:30:06 crc kubenswrapper[4733]: I0202 15:30:06.098409 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6q42z\" (UniqueName: \"kubernetes.io/projected/02261047-2e53-4b34-bde5-5da7facb243c-kube-api-access-6q42z\") pod \"openstack-galera-2\" (UID: \"02261047-2e53-4b34-bde5-5da7facb243c\") " pod="manila-kuttl-tests/openstack-galera-2" Feb 02 15:30:06 crc kubenswrapper[4733]: I0202 15:30:06.099263 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4kxk\" (UniqueName: \"kubernetes.io/projected/f177ebf9-728b-4ba2-b9b7-6c676f9d014a-kube-api-access-f4kxk\") pod \"openstack-galera-0\" (UID: \"f177ebf9-728b-4ba2-b9b7-6c676f9d014a\") " pod="manila-kuttl-tests/openstack-galera-0" Feb 02 15:30:06 crc kubenswrapper[4733]: I0202 15:30:06.105472 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xsq68\" (UniqueName: \"kubernetes.io/projected/d63f3e93-f45f-4f7d-a29b-2eaae536fa5d-kube-api-access-xsq68\") pod \"openstack-galera-1\" (UID: \"d63f3e93-f45f-4f7d-a29b-2eaae536fa5d\") " pod="manila-kuttl-tests/openstack-galera-1" Feb 02 15:30:06 crc kubenswrapper[4733]: I0202 15:30:06.120202 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/openstack-galera-0" Feb 02 15:30:06 crc kubenswrapper[4733]: I0202 15:30:06.138586 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/openstack-galera-2" Feb 02 15:30:06 crc kubenswrapper[4733]: I0202 15:30:06.150689 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/openstack-galera-1" Feb 02 15:30:06 crc kubenswrapper[4733]: I0202 15:30:06.413583 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/openstack-galera-2"] Feb 02 15:30:06 crc kubenswrapper[4733]: W0202 15:30:06.417972 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod02261047_2e53_4b34_bde5_5da7facb243c.slice/crio-31ac1ca7ba19035b1b7157aafd05dadb7f92843c69e455f557bd46e0662d7cdf WatchSource:0}: Error finding container 31ac1ca7ba19035b1b7157aafd05dadb7f92843c69e455f557bd46e0662d7cdf: Status 404 returned error can't find the container with id 31ac1ca7ba19035b1b7157aafd05dadb7f92843c69e455f557bd46e0662d7cdf Feb 02 15:30:06 crc kubenswrapper[4733]: I0202 15:30:06.459810 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/openstack-galera-1"] Feb 02 15:30:06 crc kubenswrapper[4733]: W0202 15:30:06.460946 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd63f3e93_f45f_4f7d_a29b_2eaae536fa5d.slice/crio-c9a196e8e4c248e6bcc8d38449b00c75bf0edae74116bba0e25b9abf856f4a92 WatchSource:0}: Error finding container c9a196e8e4c248e6bcc8d38449b00c75bf0edae74116bba0e25b9abf856f4a92: Status 404 returned error can't find the container with id c9a196e8e4c248e6bcc8d38449b00c75bf0edae74116bba0e25b9abf856f4a92 Feb 02 15:30:06 crc kubenswrapper[4733]: I0202 15:30:06.579124 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/openstack-galera-0"] Feb 02 15:30:06 crc kubenswrapper[4733]: W0202 15:30:06.584133 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf177ebf9_728b_4ba2_b9b7_6c676f9d014a.slice/crio-7c698d2bb4ec40a7f26727371b9ce3b5fbfc121e40731977721f74c904dbadf0 WatchSource:0}: Error finding container 7c698d2bb4ec40a7f26727371b9ce3b5fbfc121e40731977721f74c904dbadf0: Status 404 returned error can't find the container with id 7c698d2bb4ec40a7f26727371b9ce3b5fbfc121e40731977721f74c904dbadf0 Feb 02 15:30:06 crc kubenswrapper[4733]: I0202 15:30:06.670825 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/openstack-galera-0" event={"ID":"f177ebf9-728b-4ba2-b9b7-6c676f9d014a","Type":"ContainerStarted","Data":"7c698d2bb4ec40a7f26727371b9ce3b5fbfc121e40731977721f74c904dbadf0"} Feb 02 15:30:06 crc kubenswrapper[4733]: I0202 15:30:06.671989 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/openstack-galera-1" event={"ID":"d63f3e93-f45f-4f7d-a29b-2eaae536fa5d","Type":"ContainerStarted","Data":"c9a196e8e4c248e6bcc8d38449b00c75bf0edae74116bba0e25b9abf856f4a92"} Feb 02 15:30:06 crc kubenswrapper[4733]: I0202 15:30:06.673044 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/openstack-galera-2" event={"ID":"02261047-2e53-4b34-bde5-5da7facb243c","Type":"ContainerStarted","Data":"31ac1ca7ba19035b1b7157aafd05dadb7f92843c69e455f557bd46e0662d7cdf"} Feb 02 15:30:09 crc kubenswrapper[4733]: I0202 15:30:09.805275 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-b9d5f6fc5-nrtgx" Feb 02 15:30:14 crc kubenswrapper[4733]: I0202 15:30:14.651331 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["manila-kuttl-tests/memcached-0"] Feb 02 15:30:14 crc kubenswrapper[4733]: I0202 15:30:14.653466 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/memcached-0" Feb 02 15:30:14 crc kubenswrapper[4733]: I0202 15:30:14.657645 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"memcached-memcached-dockercfg-czmxl" Feb 02 15:30:14 crc kubenswrapper[4733]: I0202 15:30:14.657946 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"manila-kuttl-tests"/"memcached-config-data" Feb 02 15:30:14 crc kubenswrapper[4733]: I0202 15:30:14.667152 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/memcached-0"] Feb 02 15:30:14 crc kubenswrapper[4733]: I0202 15:30:14.690823 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e8f2d3a6-ad9f-4e1c-8f9a-41f8c7efcdac-kolla-config\") pod \"memcached-0\" (UID: \"e8f2d3a6-ad9f-4e1c-8f9a-41f8c7efcdac\") " pod="manila-kuttl-tests/memcached-0" Feb 02 15:30:14 crc kubenswrapper[4733]: I0202 15:30:14.690910 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zj2tt\" (UniqueName: \"kubernetes.io/projected/e8f2d3a6-ad9f-4e1c-8f9a-41f8c7efcdac-kube-api-access-zj2tt\") pod \"memcached-0\" (UID: \"e8f2d3a6-ad9f-4e1c-8f9a-41f8c7efcdac\") " pod="manila-kuttl-tests/memcached-0" Feb 02 15:30:14 crc kubenswrapper[4733]: I0202 15:30:14.690947 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e8f2d3a6-ad9f-4e1c-8f9a-41f8c7efcdac-config-data\") pod \"memcached-0\" (UID: \"e8f2d3a6-ad9f-4e1c-8f9a-41f8c7efcdac\") " pod="manila-kuttl-tests/memcached-0" Feb 02 15:30:14 crc kubenswrapper[4733]: I0202 15:30:14.720667 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/openstack-galera-0" event={"ID":"f177ebf9-728b-4ba2-b9b7-6c676f9d014a","Type":"ContainerStarted","Data":"6d16370fe61c34fb815b15b673736c3407ad87871969c2aab8e91a4ac281d254"} Feb 02 15:30:14 crc kubenswrapper[4733]: I0202 15:30:14.722852 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/openstack-galera-1" event={"ID":"d63f3e93-f45f-4f7d-a29b-2eaae536fa5d","Type":"ContainerStarted","Data":"cde1ae39c9ba2c79c8411fde1fac0b652cd522d574af6fd6222107e900c130c5"} Feb 02 15:30:14 crc kubenswrapper[4733]: I0202 15:30:14.724072 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/openstack-galera-2" event={"ID":"02261047-2e53-4b34-bde5-5da7facb243c","Type":"ContainerStarted","Data":"ca13066326a1d72613f501d942b75303c019bac4f9000af669ea09823a2af28b"} Feb 02 15:30:14 crc kubenswrapper[4733]: I0202 15:30:14.728121 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-vrfxd"] Feb 02 15:30:14 crc kubenswrapper[4733]: I0202 15:30:14.728843 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-vrfxd" Feb 02 15:30:14 crc kubenswrapper[4733]: I0202 15:30:14.732317 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-index-dockercfg-vtbmk" Feb 02 15:30:14 crc kubenswrapper[4733]: I0202 15:30:14.738779 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-vrfxd"] Feb 02 15:30:14 crc kubenswrapper[4733]: I0202 15:30:14.792421 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lqg7\" (UniqueName: \"kubernetes.io/projected/ab733d19-3903-4210-be9e-78bf482d848a-kube-api-access-2lqg7\") pod \"rabbitmq-cluster-operator-index-vrfxd\" (UID: \"ab733d19-3903-4210-be9e-78bf482d848a\") " pod="openstack-operators/rabbitmq-cluster-operator-index-vrfxd" Feb 02 15:30:14 crc kubenswrapper[4733]: I0202 15:30:14.792520 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zj2tt\" (UniqueName: \"kubernetes.io/projected/e8f2d3a6-ad9f-4e1c-8f9a-41f8c7efcdac-kube-api-access-zj2tt\") pod \"memcached-0\" (UID: \"e8f2d3a6-ad9f-4e1c-8f9a-41f8c7efcdac\") " pod="manila-kuttl-tests/memcached-0" Feb 02 15:30:14 crc kubenswrapper[4733]: I0202 15:30:14.792589 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e8f2d3a6-ad9f-4e1c-8f9a-41f8c7efcdac-config-data\") pod \"memcached-0\" (UID: \"e8f2d3a6-ad9f-4e1c-8f9a-41f8c7efcdac\") " pod="manila-kuttl-tests/memcached-0" Feb 02 15:30:14 crc kubenswrapper[4733]: I0202 15:30:14.792634 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e8f2d3a6-ad9f-4e1c-8f9a-41f8c7efcdac-kolla-config\") pod \"memcached-0\" (UID: \"e8f2d3a6-ad9f-4e1c-8f9a-41f8c7efcdac\") " pod="manila-kuttl-tests/memcached-0" Feb 02 15:30:14 crc kubenswrapper[4733]: I0202 15:30:14.793954 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e8f2d3a6-ad9f-4e1c-8f9a-41f8c7efcdac-kolla-config\") pod \"memcached-0\" (UID: \"e8f2d3a6-ad9f-4e1c-8f9a-41f8c7efcdac\") " pod="manila-kuttl-tests/memcached-0" Feb 02 15:30:14 crc kubenswrapper[4733]: I0202 15:30:14.808468 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e8f2d3a6-ad9f-4e1c-8f9a-41f8c7efcdac-config-data\") pod \"memcached-0\" (UID: \"e8f2d3a6-ad9f-4e1c-8f9a-41f8c7efcdac\") " pod="manila-kuttl-tests/memcached-0" Feb 02 15:30:14 crc kubenswrapper[4733]: I0202 15:30:14.820845 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zj2tt\" (UniqueName: \"kubernetes.io/projected/e8f2d3a6-ad9f-4e1c-8f9a-41f8c7efcdac-kube-api-access-zj2tt\") pod \"memcached-0\" (UID: \"e8f2d3a6-ad9f-4e1c-8f9a-41f8c7efcdac\") " pod="manila-kuttl-tests/memcached-0" Feb 02 15:30:14 crc kubenswrapper[4733]: I0202 15:30:14.894095 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lqg7\" (UniqueName: \"kubernetes.io/projected/ab733d19-3903-4210-be9e-78bf482d848a-kube-api-access-2lqg7\") pod \"rabbitmq-cluster-operator-index-vrfxd\" (UID: \"ab733d19-3903-4210-be9e-78bf482d848a\") " pod="openstack-operators/rabbitmq-cluster-operator-index-vrfxd" Feb 02 15:30:14 crc kubenswrapper[4733]: I0202 15:30:14.911180 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lqg7\" (UniqueName: \"kubernetes.io/projected/ab733d19-3903-4210-be9e-78bf482d848a-kube-api-access-2lqg7\") pod \"rabbitmq-cluster-operator-index-vrfxd\" (UID: \"ab733d19-3903-4210-be9e-78bf482d848a\") " pod="openstack-operators/rabbitmq-cluster-operator-index-vrfxd" Feb 02 15:30:14 crc kubenswrapper[4733]: I0202 15:30:14.972470 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/memcached-0" Feb 02 15:30:15 crc kubenswrapper[4733]: I0202 15:30:15.041123 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-vrfxd" Feb 02 15:30:15 crc kubenswrapper[4733]: I0202 15:30:15.444370 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/memcached-0"] Feb 02 15:30:15 crc kubenswrapper[4733]: W0202 15:30:15.445411 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode8f2d3a6_ad9f_4e1c_8f9a_41f8c7efcdac.slice/crio-4f75aaa4c7a7096a3bad8398b01deb8d12d4e519d6c78295f64cd3f241439ba6 WatchSource:0}: Error finding container 4f75aaa4c7a7096a3bad8398b01deb8d12d4e519d6c78295f64cd3f241439ba6: Status 404 returned error can't find the container with id 4f75aaa4c7a7096a3bad8398b01deb8d12d4e519d6c78295f64cd3f241439ba6 Feb 02 15:30:15 crc kubenswrapper[4733]: I0202 15:30:15.618260 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-vrfxd"] Feb 02 15:30:15 crc kubenswrapper[4733]: W0202 15:30:15.620941 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podab733d19_3903_4210_be9e_78bf482d848a.slice/crio-e210d484e29a93ef4fead5ac8a6dc381a710aa44d53c5cc5fbfd37806efe74ec WatchSource:0}: Error finding container e210d484e29a93ef4fead5ac8a6dc381a710aa44d53c5cc5fbfd37806efe74ec: Status 404 returned error can't find the container with id e210d484e29a93ef4fead5ac8a6dc381a710aa44d53c5cc5fbfd37806efe74ec Feb 02 15:30:15 crc kubenswrapper[4733]: I0202 15:30:15.729345 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/memcached-0" event={"ID":"e8f2d3a6-ad9f-4e1c-8f9a-41f8c7efcdac","Type":"ContainerStarted","Data":"4f75aaa4c7a7096a3bad8398b01deb8d12d4e519d6c78295f64cd3f241439ba6"} Feb 02 15:30:15 crc kubenswrapper[4733]: I0202 15:30:15.730895 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-vrfxd" event={"ID":"ab733d19-3903-4210-be9e-78bf482d848a","Type":"ContainerStarted","Data":"e210d484e29a93ef4fead5ac8a6dc381a710aa44d53c5cc5fbfd37806efe74ec"} Feb 02 15:30:18 crc kubenswrapper[4733]: I0202 15:30:18.798823 4733 generic.go:334] "Generic (PLEG): container finished" podID="02261047-2e53-4b34-bde5-5da7facb243c" containerID="ca13066326a1d72613f501d942b75303c019bac4f9000af669ea09823a2af28b" exitCode=0 Feb 02 15:30:18 crc kubenswrapper[4733]: I0202 15:30:18.798981 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/openstack-galera-2" event={"ID":"02261047-2e53-4b34-bde5-5da7facb243c","Type":"ContainerDied","Data":"ca13066326a1d72613f501d942b75303c019bac4f9000af669ea09823a2af28b"} Feb 02 15:30:18 crc kubenswrapper[4733]: I0202 15:30:18.915828 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-vrfxd"] Feb 02 15:30:19 crc kubenswrapper[4733]: I0202 15:30:19.536480 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-nrbdq"] Feb 02 15:30:19 crc kubenswrapper[4733]: I0202 15:30:19.537685 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-nrbdq" Feb 02 15:30:19 crc kubenswrapper[4733]: I0202 15:30:19.543142 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-nrbdq"] Feb 02 15:30:19 crc kubenswrapper[4733]: I0202 15:30:19.671009 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lc77f\" (UniqueName: \"kubernetes.io/projected/855c71f2-a025-49cb-84c8-0206b2ec7e08-kube-api-access-lc77f\") pod \"rabbitmq-cluster-operator-index-nrbdq\" (UID: \"855c71f2-a025-49cb-84c8-0206b2ec7e08\") " pod="openstack-operators/rabbitmq-cluster-operator-index-nrbdq" Feb 02 15:30:19 crc kubenswrapper[4733]: I0202 15:30:19.772343 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lc77f\" (UniqueName: \"kubernetes.io/projected/855c71f2-a025-49cb-84c8-0206b2ec7e08-kube-api-access-lc77f\") pod \"rabbitmq-cluster-operator-index-nrbdq\" (UID: \"855c71f2-a025-49cb-84c8-0206b2ec7e08\") " pod="openstack-operators/rabbitmq-cluster-operator-index-nrbdq" Feb 02 15:30:19 crc kubenswrapper[4733]: I0202 15:30:19.801981 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lc77f\" (UniqueName: \"kubernetes.io/projected/855c71f2-a025-49cb-84c8-0206b2ec7e08-kube-api-access-lc77f\") pod \"rabbitmq-cluster-operator-index-nrbdq\" (UID: \"855c71f2-a025-49cb-84c8-0206b2ec7e08\") " pod="openstack-operators/rabbitmq-cluster-operator-index-nrbdq" Feb 02 15:30:19 crc kubenswrapper[4733]: I0202 15:30:19.820680 4733 generic.go:334] "Generic (PLEG): container finished" podID="d63f3e93-f45f-4f7d-a29b-2eaae536fa5d" containerID="cde1ae39c9ba2c79c8411fde1fac0b652cd522d574af6fd6222107e900c130c5" exitCode=0 Feb 02 15:30:19 crc kubenswrapper[4733]: I0202 15:30:19.820769 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/openstack-galera-1" event={"ID":"d63f3e93-f45f-4f7d-a29b-2eaae536fa5d","Type":"ContainerDied","Data":"cde1ae39c9ba2c79c8411fde1fac0b652cd522d574af6fd6222107e900c130c5"} Feb 02 15:30:19 crc kubenswrapper[4733]: I0202 15:30:19.824834 4733 generic.go:334] "Generic (PLEG): container finished" podID="f177ebf9-728b-4ba2-b9b7-6c676f9d014a" containerID="6d16370fe61c34fb815b15b673736c3407ad87871969c2aab8e91a4ac281d254" exitCode=0 Feb 02 15:30:19 crc kubenswrapper[4733]: I0202 15:30:19.824868 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/openstack-galera-0" event={"ID":"f177ebf9-728b-4ba2-b9b7-6c676f9d014a","Type":"ContainerDied","Data":"6d16370fe61c34fb815b15b673736c3407ad87871969c2aab8e91a4ac281d254"} Feb 02 15:30:19 crc kubenswrapper[4733]: I0202 15:30:19.864064 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-nrbdq" Feb 02 15:30:20 crc kubenswrapper[4733]: I0202 15:30:20.836321 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/openstack-galera-0" event={"ID":"f177ebf9-728b-4ba2-b9b7-6c676f9d014a","Type":"ContainerStarted","Data":"69b15a1f6699ebd76df18a6d10489a9a8f470c2c56ccab52ad3c08a533bb0a92"} Feb 02 15:30:20 crc kubenswrapper[4733]: I0202 15:30:20.837397 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-nrbdq"] Feb 02 15:30:20 crc kubenswrapper[4733]: W0202 15:30:20.838787 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod855c71f2_a025_49cb_84c8_0206b2ec7e08.slice/crio-6345b9f9c8b0745137ab98048d92f341bfd5f9f90ae157501f391a8143376b1b WatchSource:0}: Error finding container 6345b9f9c8b0745137ab98048d92f341bfd5f9f90ae157501f391a8143376b1b: Status 404 returned error can't find the container with id 6345b9f9c8b0745137ab98048d92f341bfd5f9f90ae157501f391a8143376b1b Feb 02 15:30:20 crc kubenswrapper[4733]: I0202 15:30:20.839180 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/memcached-0" event={"ID":"e8f2d3a6-ad9f-4e1c-8f9a-41f8c7efcdac","Type":"ContainerStarted","Data":"60f342505fd409f05a577cf0a9883bd43822c387690a1b4617f376e8acec2dc4"} Feb 02 15:30:20 crc kubenswrapper[4733]: I0202 15:30:20.839236 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="manila-kuttl-tests/memcached-0" Feb 02 15:30:20 crc kubenswrapper[4733]: I0202 15:30:20.854219 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/rabbitmq-cluster-operator-index-vrfxd" podUID="ab733d19-3903-4210-be9e-78bf482d848a" containerName="registry-server" containerID="cri-o://d62e8e918e498c45c7e38987547a8950d51d5dfc7b0448996718c42698dfa7c7" gracePeriod=2 Feb 02 15:30:20 crc kubenswrapper[4733]: I0202 15:30:20.854359 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-vrfxd" event={"ID":"ab733d19-3903-4210-be9e-78bf482d848a","Type":"ContainerStarted","Data":"d62e8e918e498c45c7e38987547a8950d51d5dfc7b0448996718c42698dfa7c7"} Feb 02 15:30:20 crc kubenswrapper[4733]: I0202 15:30:20.860844 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/openstack-galera-1" event={"ID":"d63f3e93-f45f-4f7d-a29b-2eaae536fa5d","Type":"ContainerStarted","Data":"77d577f4756bae43064e83c2d18077bad1b459a16e5122f3a5125f7ac44f7514"} Feb 02 15:30:20 crc kubenswrapper[4733]: I0202 15:30:20.870309 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="manila-kuttl-tests/openstack-galera-0" podStartSLOduration=9.235896476 podStartE2EDuration="16.87028867s" podCreationTimestamp="2026-02-02 15:30:04 +0000 UTC" firstStartedPulling="2026-02-02 15:30:06.586071957 +0000 UTC m=+890.037533315" lastFinishedPulling="2026-02-02 15:30:14.220464161 +0000 UTC m=+897.671925509" observedRunningTime="2026-02-02 15:30:20.867947809 +0000 UTC m=+904.319409197" watchObservedRunningTime="2026-02-02 15:30:20.87028867 +0000 UTC m=+904.321750028" Feb 02 15:30:20 crc kubenswrapper[4733]: I0202 15:30:20.870807 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/openstack-galera-2" event={"ID":"02261047-2e53-4b34-bde5-5da7facb243c","Type":"ContainerStarted","Data":"8a9cc594f867e17e58e08d3984767170690347ae2f35b2e53796aa7c6d0db5f2"} Feb 02 15:30:20 crc kubenswrapper[4733]: I0202 15:30:20.902263 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="manila-kuttl-tests/openstack-galera-1" podStartSLOduration=9.125722023 podStartE2EDuration="16.902242824s" podCreationTimestamp="2026-02-02 15:30:04 +0000 UTC" firstStartedPulling="2026-02-02 15:30:06.463219432 +0000 UTC m=+889.914680790" lastFinishedPulling="2026-02-02 15:30:14.239740223 +0000 UTC m=+897.691201591" observedRunningTime="2026-02-02 15:30:20.897460939 +0000 UTC m=+904.348922317" watchObservedRunningTime="2026-02-02 15:30:20.902242824 +0000 UTC m=+904.353704182" Feb 02 15:30:20 crc kubenswrapper[4733]: I0202 15:30:20.919430 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-index-vrfxd" podStartSLOduration=2.107976158 podStartE2EDuration="6.919411542s" podCreationTimestamp="2026-02-02 15:30:14 +0000 UTC" firstStartedPulling="2026-02-02 15:30:15.623052568 +0000 UTC m=+899.074513926" lastFinishedPulling="2026-02-02 15:30:20.434487932 +0000 UTC m=+903.885949310" observedRunningTime="2026-02-02 15:30:20.91435778 +0000 UTC m=+904.365819158" watchObservedRunningTime="2026-02-02 15:30:20.919411542 +0000 UTC m=+904.370872900" Feb 02 15:30:20 crc kubenswrapper[4733]: I0202 15:30:20.955173 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="manila-kuttl-tests/openstack-galera-2" podStartSLOduration=9.093255255 podStartE2EDuration="16.955138114s" podCreationTimestamp="2026-02-02 15:30:04 +0000 UTC" firstStartedPulling="2026-02-02 15:30:06.420412235 +0000 UTC m=+889.871873593" lastFinishedPulling="2026-02-02 15:30:14.282295094 +0000 UTC m=+897.733756452" observedRunningTime="2026-02-02 15:30:20.952264509 +0000 UTC m=+904.403725877" watchObservedRunningTime="2026-02-02 15:30:20.955138114 +0000 UTC m=+904.406599472" Feb 02 15:30:20 crc kubenswrapper[4733]: I0202 15:30:20.956207 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="manila-kuttl-tests/memcached-0" podStartSLOduration=3.273163734 podStartE2EDuration="6.956201282s" podCreationTimestamp="2026-02-02 15:30:14 +0000 UTC" firstStartedPulling="2026-02-02 15:30:15.447486658 +0000 UTC m=+898.898948016" lastFinishedPulling="2026-02-02 15:30:19.130524206 +0000 UTC m=+902.581985564" observedRunningTime="2026-02-02 15:30:20.934703401 +0000 UTC m=+904.386164769" watchObservedRunningTime="2026-02-02 15:30:20.956201282 +0000 UTC m=+904.407662640" Feb 02 15:30:21 crc kubenswrapper[4733]: I0202 15:30:21.204923 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-vrfxd" Feb 02 15:30:21 crc kubenswrapper[4733]: I0202 15:30:21.309440 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2lqg7\" (UniqueName: \"kubernetes.io/projected/ab733d19-3903-4210-be9e-78bf482d848a-kube-api-access-2lqg7\") pod \"ab733d19-3903-4210-be9e-78bf482d848a\" (UID: \"ab733d19-3903-4210-be9e-78bf482d848a\") " Feb 02 15:30:21 crc kubenswrapper[4733]: I0202 15:30:21.315301 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab733d19-3903-4210-be9e-78bf482d848a-kube-api-access-2lqg7" (OuterVolumeSpecName: "kube-api-access-2lqg7") pod "ab733d19-3903-4210-be9e-78bf482d848a" (UID: "ab733d19-3903-4210-be9e-78bf482d848a"). InnerVolumeSpecName "kube-api-access-2lqg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:30:21 crc kubenswrapper[4733]: I0202 15:30:21.410816 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2lqg7\" (UniqueName: \"kubernetes.io/projected/ab733d19-3903-4210-be9e-78bf482d848a-kube-api-access-2lqg7\") on node \"crc\" DevicePath \"\"" Feb 02 15:30:21 crc kubenswrapper[4733]: I0202 15:30:21.879477 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-nrbdq" event={"ID":"855c71f2-a025-49cb-84c8-0206b2ec7e08","Type":"ContainerStarted","Data":"ba22d52e2c88845b48608d81df18d063e6040d2d90a6f634187534bf7ca77329"} Feb 02 15:30:21 crc kubenswrapper[4733]: I0202 15:30:21.879802 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-nrbdq" event={"ID":"855c71f2-a025-49cb-84c8-0206b2ec7e08","Type":"ContainerStarted","Data":"6345b9f9c8b0745137ab98048d92f341bfd5f9f90ae157501f391a8143376b1b"} Feb 02 15:30:21 crc kubenswrapper[4733]: I0202 15:30:21.883576 4733 generic.go:334] "Generic (PLEG): container finished" podID="ab733d19-3903-4210-be9e-78bf482d848a" containerID="d62e8e918e498c45c7e38987547a8950d51d5dfc7b0448996718c42698dfa7c7" exitCode=0 Feb 02 15:30:21 crc kubenswrapper[4733]: I0202 15:30:21.883616 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-vrfxd" Feb 02 15:30:21 crc kubenswrapper[4733]: I0202 15:30:21.883675 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-vrfxd" event={"ID":"ab733d19-3903-4210-be9e-78bf482d848a","Type":"ContainerDied","Data":"d62e8e918e498c45c7e38987547a8950d51d5dfc7b0448996718c42698dfa7c7"} Feb 02 15:30:21 crc kubenswrapper[4733]: I0202 15:30:21.883704 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-vrfxd" event={"ID":"ab733d19-3903-4210-be9e-78bf482d848a","Type":"ContainerDied","Data":"e210d484e29a93ef4fead5ac8a6dc381a710aa44d53c5cc5fbfd37806efe74ec"} Feb 02 15:30:21 crc kubenswrapper[4733]: I0202 15:30:21.883724 4733 scope.go:117] "RemoveContainer" containerID="d62e8e918e498c45c7e38987547a8950d51d5dfc7b0448996718c42698dfa7c7" Feb 02 15:30:21 crc kubenswrapper[4733]: I0202 15:30:21.898939 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-index-nrbdq" podStartSLOduration=2.417123025 podStartE2EDuration="2.898916704s" podCreationTimestamp="2026-02-02 15:30:19 +0000 UTC" firstStartedPulling="2026-02-02 15:30:20.847098045 +0000 UTC m=+904.298559403" lastFinishedPulling="2026-02-02 15:30:21.328891724 +0000 UTC m=+904.780353082" observedRunningTime="2026-02-02 15:30:21.894018706 +0000 UTC m=+905.345480074" watchObservedRunningTime="2026-02-02 15:30:21.898916704 +0000 UTC m=+905.350378062" Feb 02 15:30:21 crc kubenswrapper[4733]: I0202 15:30:21.916069 4733 scope.go:117] "RemoveContainer" containerID="d62e8e918e498c45c7e38987547a8950d51d5dfc7b0448996718c42698dfa7c7" Feb 02 15:30:21 crc kubenswrapper[4733]: E0202 15:30:21.916611 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d62e8e918e498c45c7e38987547a8950d51d5dfc7b0448996718c42698dfa7c7\": container with ID starting with d62e8e918e498c45c7e38987547a8950d51d5dfc7b0448996718c42698dfa7c7 not found: ID does not exist" containerID="d62e8e918e498c45c7e38987547a8950d51d5dfc7b0448996718c42698dfa7c7" Feb 02 15:30:21 crc kubenswrapper[4733]: I0202 15:30:21.916656 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d62e8e918e498c45c7e38987547a8950d51d5dfc7b0448996718c42698dfa7c7"} err="failed to get container status \"d62e8e918e498c45c7e38987547a8950d51d5dfc7b0448996718c42698dfa7c7\": rpc error: code = NotFound desc = could not find container \"d62e8e918e498c45c7e38987547a8950d51d5dfc7b0448996718c42698dfa7c7\": container with ID starting with d62e8e918e498c45c7e38987547a8950d51d5dfc7b0448996718c42698dfa7c7 not found: ID does not exist" Feb 02 15:30:21 crc kubenswrapper[4733]: I0202 15:30:21.925855 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-vrfxd"] Feb 02 15:30:21 crc kubenswrapper[4733]: I0202 15:30:21.934621 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-vrfxd"] Feb 02 15:30:23 crc kubenswrapper[4733]: I0202 15:30:23.267766 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab733d19-3903-4210-be9e-78bf482d848a" path="/var/lib/kubelet/pods/ab733d19-3903-4210-be9e-78bf482d848a/volumes" Feb 02 15:30:26 crc kubenswrapper[4733]: I0202 15:30:26.121224 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="manila-kuttl-tests/openstack-galera-0" Feb 02 15:30:26 crc kubenswrapper[4733]: I0202 15:30:26.121550 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="manila-kuttl-tests/openstack-galera-0" Feb 02 15:30:26 crc kubenswrapper[4733]: I0202 15:30:26.139204 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="manila-kuttl-tests/openstack-galera-2" Feb 02 15:30:26 crc kubenswrapper[4733]: I0202 15:30:26.139446 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="manila-kuttl-tests/openstack-galera-2" Feb 02 15:30:26 crc kubenswrapper[4733]: I0202 15:30:26.150945 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="manila-kuttl-tests/openstack-galera-1" Feb 02 15:30:26 crc kubenswrapper[4733]: I0202 15:30:26.151032 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="manila-kuttl-tests/openstack-galera-1" Feb 02 15:30:29 crc kubenswrapper[4733]: I0202 15:30:29.865708 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/rabbitmq-cluster-operator-index-nrbdq" Feb 02 15:30:29 crc kubenswrapper[4733]: I0202 15:30:29.866310 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/rabbitmq-cluster-operator-index-nrbdq" Feb 02 15:30:29 crc kubenswrapper[4733]: I0202 15:30:29.892221 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/rabbitmq-cluster-operator-index-nrbdq" Feb 02 15:30:29 crc kubenswrapper[4733]: I0202 15:30:29.973792 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="manila-kuttl-tests/memcached-0" Feb 02 15:30:29 crc kubenswrapper[4733]: I0202 15:30:29.974027 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/rabbitmq-cluster-operator-index-nrbdq" Feb 02 15:30:30 crc kubenswrapper[4733]: I0202 15:30:30.794803 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="manila-kuttl-tests/openstack-galera-2" Feb 02 15:30:30 crc kubenswrapper[4733]: I0202 15:30:30.895689 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="manila-kuttl-tests/openstack-galera-2" Feb 02 15:30:34 crc kubenswrapper[4733]: I0202 15:30:34.871103 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["manila-kuttl-tests/root-account-create-update-4jtql"] Feb 02 15:30:34 crc kubenswrapper[4733]: E0202 15:30:34.872006 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab733d19-3903-4210-be9e-78bf482d848a" containerName="registry-server" Feb 02 15:30:34 crc kubenswrapper[4733]: I0202 15:30:34.872022 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab733d19-3903-4210-be9e-78bf482d848a" containerName="registry-server" Feb 02 15:30:34 crc kubenswrapper[4733]: I0202 15:30:34.872166 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab733d19-3903-4210-be9e-78bf482d848a" containerName="registry-server" Feb 02 15:30:34 crc kubenswrapper[4733]: I0202 15:30:34.872689 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/root-account-create-update-4jtql" Feb 02 15:30:34 crc kubenswrapper[4733]: I0202 15:30:34.874680 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"openstack-mariadb-root-db-secret" Feb 02 15:30:34 crc kubenswrapper[4733]: I0202 15:30:34.882869 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/root-account-create-update-4jtql"] Feb 02 15:30:34 crc kubenswrapper[4733]: I0202 15:30:34.900278 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f36bcd25-3798-4fbf-b317-b6d52b53af05-operator-scripts\") pod \"root-account-create-update-4jtql\" (UID: \"f36bcd25-3798-4fbf-b317-b6d52b53af05\") " pod="manila-kuttl-tests/root-account-create-update-4jtql" Feb 02 15:30:34 crc kubenswrapper[4733]: I0202 15:30:34.900330 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-899jr\" (UniqueName: \"kubernetes.io/projected/f36bcd25-3798-4fbf-b317-b6d52b53af05-kube-api-access-899jr\") pod \"root-account-create-update-4jtql\" (UID: \"f36bcd25-3798-4fbf-b317-b6d52b53af05\") " pod="manila-kuttl-tests/root-account-create-update-4jtql" Feb 02 15:30:35 crc kubenswrapper[4733]: I0202 15:30:35.002020 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f36bcd25-3798-4fbf-b317-b6d52b53af05-operator-scripts\") pod \"root-account-create-update-4jtql\" (UID: \"f36bcd25-3798-4fbf-b317-b6d52b53af05\") " pod="manila-kuttl-tests/root-account-create-update-4jtql" Feb 02 15:30:35 crc kubenswrapper[4733]: I0202 15:30:35.002079 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-899jr\" (UniqueName: \"kubernetes.io/projected/f36bcd25-3798-4fbf-b317-b6d52b53af05-kube-api-access-899jr\") pod \"root-account-create-update-4jtql\" (UID: \"f36bcd25-3798-4fbf-b317-b6d52b53af05\") " pod="manila-kuttl-tests/root-account-create-update-4jtql" Feb 02 15:30:35 crc kubenswrapper[4733]: I0202 15:30:35.003278 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f36bcd25-3798-4fbf-b317-b6d52b53af05-operator-scripts\") pod \"root-account-create-update-4jtql\" (UID: \"f36bcd25-3798-4fbf-b317-b6d52b53af05\") " pod="manila-kuttl-tests/root-account-create-update-4jtql" Feb 02 15:30:35 crc kubenswrapper[4733]: I0202 15:30:35.019713 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-899jr\" (UniqueName: \"kubernetes.io/projected/f36bcd25-3798-4fbf-b317-b6d52b53af05-kube-api-access-899jr\") pod \"root-account-create-update-4jtql\" (UID: \"f36bcd25-3798-4fbf-b317-b6d52b53af05\") " pod="manila-kuttl-tests/root-account-create-update-4jtql" Feb 02 15:30:35 crc kubenswrapper[4733]: I0202 15:30:35.193161 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/root-account-create-update-4jtql" Feb 02 15:30:35 crc kubenswrapper[4733]: I0202 15:30:35.684088 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/root-account-create-update-4jtql"] Feb 02 15:30:35 crc kubenswrapper[4733]: W0202 15:30:35.689692 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf36bcd25_3798_4fbf_b317_b6d52b53af05.slice/crio-96348660979cf8a8bb51d09c032cbefbd4a504c7fb798a0cf8280609f9fda987 WatchSource:0}: Error finding container 96348660979cf8a8bb51d09c032cbefbd4a504c7fb798a0cf8280609f9fda987: Status 404 returned error can't find the container with id 96348660979cf8a8bb51d09c032cbefbd4a504c7fb798a0cf8280609f9fda987 Feb 02 15:30:35 crc kubenswrapper[4733]: I0202 15:30:35.978535 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/root-account-create-update-4jtql" event={"ID":"f36bcd25-3798-4fbf-b317-b6d52b53af05","Type":"ContainerStarted","Data":"02a2e1727447b5eada81717bc18a5e1fa01249afa350c42cb2d50d68ce2ee5fd"} Feb 02 15:30:35 crc kubenswrapper[4733]: I0202 15:30:35.978585 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/root-account-create-update-4jtql" event={"ID":"f36bcd25-3798-4fbf-b317-b6d52b53af05","Type":"ContainerStarted","Data":"96348660979cf8a8bb51d09c032cbefbd4a504c7fb798a0cf8280609f9fda987"} Feb 02 15:30:35 crc kubenswrapper[4733]: I0202 15:30:35.995329 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="manila-kuttl-tests/root-account-create-update-4jtql" podStartSLOduration=1.995314783 podStartE2EDuration="1.995314783s" podCreationTimestamp="2026-02-02 15:30:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:30:35.992107668 +0000 UTC m=+919.443569046" watchObservedRunningTime="2026-02-02 15:30:35.995314783 +0000 UTC m=+919.446776141" Feb 02 15:30:36 crc kubenswrapper[4733]: I0202 15:30:36.217560 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="manila-kuttl-tests/openstack-galera-2" podUID="02261047-2e53-4b34-bde5-5da7facb243c" containerName="galera" probeResult="failure" output=< Feb 02 15:30:36 crc kubenswrapper[4733]: wsrep_local_state_comment (Donor/Desynced) differs from Synced Feb 02 15:30:36 crc kubenswrapper[4733]: > Feb 02 15:30:37 crc kubenswrapper[4733]: I0202 15:30:37.996370 4733 generic.go:334] "Generic (PLEG): container finished" podID="f36bcd25-3798-4fbf-b317-b6d52b53af05" containerID="02a2e1727447b5eada81717bc18a5e1fa01249afa350c42cb2d50d68ce2ee5fd" exitCode=0 Feb 02 15:30:37 crc kubenswrapper[4733]: I0202 15:30:37.996517 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/root-account-create-update-4jtql" event={"ID":"f36bcd25-3798-4fbf-b317-b6d52b53af05","Type":"ContainerDied","Data":"02a2e1727447b5eada81717bc18a5e1fa01249afa350c42cb2d50d68ce2ee5fd"} Feb 02 15:30:38 crc kubenswrapper[4733]: I0202 15:30:38.769944 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590wfw27"] Feb 02 15:30:38 crc kubenswrapper[4733]: I0202 15:30:38.771338 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590wfw27" Feb 02 15:30:38 crc kubenswrapper[4733]: I0202 15:30:38.774070 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-2mtp8" Feb 02 15:30:38 crc kubenswrapper[4733]: I0202 15:30:38.788161 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590wfw27"] Feb 02 15:30:38 crc kubenswrapper[4733]: I0202 15:30:38.859991 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590wfw27\" (UID: \"5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590wfw27" Feb 02 15:30:38 crc kubenswrapper[4733]: I0202 15:30:38.860060 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wwnc\" (UniqueName: \"kubernetes.io/projected/5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6-kube-api-access-4wwnc\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590wfw27\" (UID: \"5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590wfw27" Feb 02 15:30:38 crc kubenswrapper[4733]: I0202 15:30:38.860119 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590wfw27\" (UID: \"5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590wfw27" Feb 02 15:30:38 crc kubenswrapper[4733]: I0202 15:30:38.961579 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590wfw27\" (UID: \"5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590wfw27" Feb 02 15:30:38 crc kubenswrapper[4733]: I0202 15:30:38.961661 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590wfw27\" (UID: \"5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590wfw27" Feb 02 15:30:38 crc kubenswrapper[4733]: I0202 15:30:38.961718 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wwnc\" (UniqueName: \"kubernetes.io/projected/5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6-kube-api-access-4wwnc\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590wfw27\" (UID: \"5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590wfw27" Feb 02 15:30:38 crc kubenswrapper[4733]: I0202 15:30:38.962617 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590wfw27\" (UID: \"5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590wfw27" Feb 02 15:30:38 crc kubenswrapper[4733]: I0202 15:30:38.962940 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590wfw27\" (UID: \"5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590wfw27" Feb 02 15:30:38 crc kubenswrapper[4733]: I0202 15:30:38.982225 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wwnc\" (UniqueName: \"kubernetes.io/projected/5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6-kube-api-access-4wwnc\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590wfw27\" (UID: \"5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590wfw27" Feb 02 15:30:39 crc kubenswrapper[4733]: I0202 15:30:39.095456 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590wfw27" Feb 02 15:30:39 crc kubenswrapper[4733]: I0202 15:30:39.338453 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/root-account-create-update-4jtql" Feb 02 15:30:39 crc kubenswrapper[4733]: I0202 15:30:39.368041 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f36bcd25-3798-4fbf-b317-b6d52b53af05-operator-scripts\") pod \"f36bcd25-3798-4fbf-b317-b6d52b53af05\" (UID: \"f36bcd25-3798-4fbf-b317-b6d52b53af05\") " Feb 02 15:30:39 crc kubenswrapper[4733]: I0202 15:30:39.368241 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-899jr\" (UniqueName: \"kubernetes.io/projected/f36bcd25-3798-4fbf-b317-b6d52b53af05-kube-api-access-899jr\") pod \"f36bcd25-3798-4fbf-b317-b6d52b53af05\" (UID: \"f36bcd25-3798-4fbf-b317-b6d52b53af05\") " Feb 02 15:30:39 crc kubenswrapper[4733]: I0202 15:30:39.368848 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f36bcd25-3798-4fbf-b317-b6d52b53af05-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f36bcd25-3798-4fbf-b317-b6d52b53af05" (UID: "f36bcd25-3798-4fbf-b317-b6d52b53af05"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:30:39 crc kubenswrapper[4733]: I0202 15:30:39.372414 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f36bcd25-3798-4fbf-b317-b6d52b53af05-kube-api-access-899jr" (OuterVolumeSpecName: "kube-api-access-899jr") pod "f36bcd25-3798-4fbf-b317-b6d52b53af05" (UID: "f36bcd25-3798-4fbf-b317-b6d52b53af05"). InnerVolumeSpecName "kube-api-access-899jr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:30:39 crc kubenswrapper[4733]: I0202 15:30:39.470410 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-899jr\" (UniqueName: \"kubernetes.io/projected/f36bcd25-3798-4fbf-b317-b6d52b53af05-kube-api-access-899jr\") on node \"crc\" DevicePath \"\"" Feb 02 15:30:39 crc kubenswrapper[4733]: I0202 15:30:39.470448 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f36bcd25-3798-4fbf-b317-b6d52b53af05-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 15:30:39 crc kubenswrapper[4733]: I0202 15:30:39.555510 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590wfw27"] Feb 02 15:30:39 crc kubenswrapper[4733]: W0202 15:30:39.558619 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5cc5b7d3_4371_4f14_ae5b_712db7f1d2a6.slice/crio-966c20e7bdb107a964806551c46b9dbc246daa349df82f99b9629b9019190c64 WatchSource:0}: Error finding container 966c20e7bdb107a964806551c46b9dbc246daa349df82f99b9629b9019190c64: Status 404 returned error can't find the container with id 966c20e7bdb107a964806551c46b9dbc246daa349df82f99b9629b9019190c64 Feb 02 15:30:40 crc kubenswrapper[4733]: I0202 15:30:40.010343 4733 generic.go:334] "Generic (PLEG): container finished" podID="5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6" containerID="a3d0ca21093d66d012f1a132f486134fba8c350e499831e28e259fd9ad01f1ce" exitCode=0 Feb 02 15:30:40 crc kubenswrapper[4733]: I0202 15:30:40.010406 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590wfw27" event={"ID":"5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6","Type":"ContainerDied","Data":"a3d0ca21093d66d012f1a132f486134fba8c350e499831e28e259fd9ad01f1ce"} Feb 02 15:30:40 crc kubenswrapper[4733]: I0202 15:30:40.010693 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590wfw27" event={"ID":"5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6","Type":"ContainerStarted","Data":"966c20e7bdb107a964806551c46b9dbc246daa349df82f99b9629b9019190c64"} Feb 02 15:30:40 crc kubenswrapper[4733]: I0202 15:30:40.013510 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/root-account-create-update-4jtql" event={"ID":"f36bcd25-3798-4fbf-b317-b6d52b53af05","Type":"ContainerDied","Data":"96348660979cf8a8bb51d09c032cbefbd4a504c7fb798a0cf8280609f9fda987"} Feb 02 15:30:40 crc kubenswrapper[4733]: I0202 15:30:40.013538 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="96348660979cf8a8bb51d09c032cbefbd4a504c7fb798a0cf8280609f9fda987" Feb 02 15:30:40 crc kubenswrapper[4733]: I0202 15:30:40.013555 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/root-account-create-update-4jtql" Feb 02 15:30:41 crc kubenswrapper[4733]: I0202 15:30:41.020136 4733 generic.go:334] "Generic (PLEG): container finished" podID="5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6" containerID="a7d14acc7c8c90a61ee49da51a9d21cecb725ad53d1364bd144b14b64432aae5" exitCode=0 Feb 02 15:30:41 crc kubenswrapper[4733]: I0202 15:30:41.020196 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590wfw27" event={"ID":"5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6","Type":"ContainerDied","Data":"a7d14acc7c8c90a61ee49da51a9d21cecb725ad53d1364bd144b14b64432aae5"} Feb 02 15:30:41 crc kubenswrapper[4733]: I0202 15:30:41.549845 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="manila-kuttl-tests/openstack-galera-0" Feb 02 15:30:41 crc kubenswrapper[4733]: I0202 15:30:41.637245 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="manila-kuttl-tests/openstack-galera-0" Feb 02 15:30:42 crc kubenswrapper[4733]: I0202 15:30:42.030537 4733 generic.go:334] "Generic (PLEG): container finished" podID="5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6" containerID="fdb7a4d2bece5b4300fa82ba55f8f2bb9a7e357f8b4066781a7982091f253779" exitCode=0 Feb 02 15:30:42 crc kubenswrapper[4733]: I0202 15:30:42.030698 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590wfw27" event={"ID":"5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6","Type":"ContainerDied","Data":"fdb7a4d2bece5b4300fa82ba55f8f2bb9a7e357f8b4066781a7982091f253779"} Feb 02 15:30:43 crc kubenswrapper[4733]: I0202 15:30:43.359009 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590wfw27" Feb 02 15:30:43 crc kubenswrapper[4733]: I0202 15:30:43.434075 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6-util\") pod \"5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6\" (UID: \"5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6\") " Feb 02 15:30:43 crc kubenswrapper[4733]: I0202 15:30:43.434619 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6-bundle\") pod \"5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6\" (UID: \"5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6\") " Feb 02 15:30:43 crc kubenswrapper[4733]: I0202 15:30:43.434713 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4wwnc\" (UniqueName: \"kubernetes.io/projected/5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6-kube-api-access-4wwnc\") pod \"5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6\" (UID: \"5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6\") " Feb 02 15:30:43 crc kubenswrapper[4733]: I0202 15:30:43.436917 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6-bundle" (OuterVolumeSpecName: "bundle") pod "5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6" (UID: "5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:30:43 crc kubenswrapper[4733]: I0202 15:30:43.441017 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6-kube-api-access-4wwnc" (OuterVolumeSpecName: "kube-api-access-4wwnc") pod "5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6" (UID: "5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6"). InnerVolumeSpecName "kube-api-access-4wwnc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:30:43 crc kubenswrapper[4733]: I0202 15:30:43.443838 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6-util" (OuterVolumeSpecName: "util") pod "5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6" (UID: "5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:30:43 crc kubenswrapper[4733]: I0202 15:30:43.536381 4733 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6-util\") on node \"crc\" DevicePath \"\"" Feb 02 15:30:43 crc kubenswrapper[4733]: I0202 15:30:43.536415 4733 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 15:30:43 crc kubenswrapper[4733]: I0202 15:30:43.536428 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4wwnc\" (UniqueName: \"kubernetes.io/projected/5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6-kube-api-access-4wwnc\") on node \"crc\" DevicePath \"\"" Feb 02 15:30:44 crc kubenswrapper[4733]: I0202 15:30:44.045468 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590wfw27" event={"ID":"5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6","Type":"ContainerDied","Data":"966c20e7bdb107a964806551c46b9dbc246daa349df82f99b9629b9019190c64"} Feb 02 15:30:44 crc kubenswrapper[4733]: I0202 15:30:44.045773 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="966c20e7bdb107a964806551c46b9dbc246daa349df82f99b9629b9019190c64" Feb 02 15:30:44 crc kubenswrapper[4733]: I0202 15:30:44.045520 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590wfw27" Feb 02 15:30:46 crc kubenswrapper[4733]: I0202 15:30:46.043590 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="manila-kuttl-tests/openstack-galera-1" Feb 02 15:30:46 crc kubenswrapper[4733]: I0202 15:30:46.134900 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="manila-kuttl-tests/openstack-galera-1" Feb 02 15:30:50 crc kubenswrapper[4733]: I0202 15:30:50.579392 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-gzhj7"] Feb 02 15:30:50 crc kubenswrapper[4733]: E0202 15:30:50.580375 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6" containerName="util" Feb 02 15:30:50 crc kubenswrapper[4733]: I0202 15:30:50.580396 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6" containerName="util" Feb 02 15:30:50 crc kubenswrapper[4733]: E0202 15:30:50.580422 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f36bcd25-3798-4fbf-b317-b6d52b53af05" containerName="mariadb-account-create-update" Feb 02 15:30:50 crc kubenswrapper[4733]: I0202 15:30:50.580434 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f36bcd25-3798-4fbf-b317-b6d52b53af05" containerName="mariadb-account-create-update" Feb 02 15:30:50 crc kubenswrapper[4733]: E0202 15:30:50.580462 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6" containerName="pull" Feb 02 15:30:50 crc kubenswrapper[4733]: I0202 15:30:50.580473 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6" containerName="pull" Feb 02 15:30:50 crc kubenswrapper[4733]: E0202 15:30:50.580490 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6" containerName="extract" Feb 02 15:30:50 crc kubenswrapper[4733]: I0202 15:30:50.580501 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6" containerName="extract" Feb 02 15:30:50 crc kubenswrapper[4733]: I0202 15:30:50.580686 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f36bcd25-3798-4fbf-b317-b6d52b53af05" containerName="mariadb-account-create-update" Feb 02 15:30:50 crc kubenswrapper[4733]: I0202 15:30:50.580717 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6" containerName="extract" Feb 02 15:30:50 crc kubenswrapper[4733]: I0202 15:30:50.581352 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-gzhj7" Feb 02 15:30:50 crc kubenswrapper[4733]: I0202 15:30:50.585247 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-dockercfg-qms8n" Feb 02 15:30:50 crc kubenswrapper[4733]: I0202 15:30:50.592557 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-gzhj7"] Feb 02 15:30:50 crc kubenswrapper[4733]: I0202 15:30:50.634918 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrcrm\" (UniqueName: \"kubernetes.io/projected/2e4d2825-196e-4bf2-8f47-270645f2267b-kube-api-access-lrcrm\") pod \"rabbitmq-cluster-operator-779fc9694b-gzhj7\" (UID: \"2e4d2825-196e-4bf2-8f47-270645f2267b\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-gzhj7" Feb 02 15:30:50 crc kubenswrapper[4733]: I0202 15:30:50.736927 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrcrm\" (UniqueName: \"kubernetes.io/projected/2e4d2825-196e-4bf2-8f47-270645f2267b-kube-api-access-lrcrm\") pod \"rabbitmq-cluster-operator-779fc9694b-gzhj7\" (UID: \"2e4d2825-196e-4bf2-8f47-270645f2267b\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-gzhj7" Feb 02 15:30:50 crc kubenswrapper[4733]: I0202 15:30:50.781348 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrcrm\" (UniqueName: \"kubernetes.io/projected/2e4d2825-196e-4bf2-8f47-270645f2267b-kube-api-access-lrcrm\") pod \"rabbitmq-cluster-operator-779fc9694b-gzhj7\" (UID: \"2e4d2825-196e-4bf2-8f47-270645f2267b\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-gzhj7" Feb 02 15:30:50 crc kubenswrapper[4733]: I0202 15:30:50.936736 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-gzhj7" Feb 02 15:30:51 crc kubenswrapper[4733]: I0202 15:30:51.359857 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-gzhj7"] Feb 02 15:30:51 crc kubenswrapper[4733]: W0202 15:30:51.370669 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod2e4d2825_196e_4bf2_8f47_270645f2267b.slice/crio-fa50ca92d643ab3b4bda92b4db4e509f85a4a7b5d120380bb525eae0a60fd7b8 WatchSource:0}: Error finding container fa50ca92d643ab3b4bda92b4db4e509f85a4a7b5d120380bb525eae0a60fd7b8: Status 404 returned error can't find the container with id fa50ca92d643ab3b4bda92b4db4e509f85a4a7b5d120380bb525eae0a60fd7b8 Feb 02 15:30:52 crc kubenswrapper[4733]: I0202 15:30:52.103668 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-gzhj7" event={"ID":"2e4d2825-196e-4bf2-8f47-270645f2267b","Type":"ContainerStarted","Data":"fa50ca92d643ab3b4bda92b4db4e509f85a4a7b5d120380bb525eae0a60fd7b8"} Feb 02 15:30:55 crc kubenswrapper[4733]: I0202 15:30:55.134575 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-gzhj7" event={"ID":"2e4d2825-196e-4bf2-8f47-270645f2267b","Type":"ContainerStarted","Data":"99ee0047c1f019d60cd3d61bf905abf882cb05edbb2b3e61679da900a5465bc4"} Feb 02 15:30:55 crc kubenswrapper[4733]: I0202 15:30:55.166233 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-gzhj7" podStartSLOduration=2.1927624 podStartE2EDuration="5.166207388s" podCreationTimestamp="2026-02-02 15:30:50 +0000 UTC" firstStartedPulling="2026-02-02 15:30:51.375693158 +0000 UTC m=+934.827154526" lastFinishedPulling="2026-02-02 15:30:54.349138156 +0000 UTC m=+937.800599514" observedRunningTime="2026-02-02 15:30:55.161448102 +0000 UTC m=+938.612909500" watchObservedRunningTime="2026-02-02 15:30:55.166207388 +0000 UTC m=+938.617668776" Feb 02 15:30:58 crc kubenswrapper[4733]: I0202 15:30:58.527010 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-index-6wnsf"] Feb 02 15:30:58 crc kubenswrapper[4733]: I0202 15:30:58.528218 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-6wnsf" Feb 02 15:30:58 crc kubenswrapper[4733]: I0202 15:30:58.534490 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-index-dockercfg-ftwq9" Feb 02 15:30:58 crc kubenswrapper[4733]: I0202 15:30:58.542932 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-6wnsf"] Feb 02 15:30:58 crc kubenswrapper[4733]: I0202 15:30:58.649853 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wb95q\" (UniqueName: \"kubernetes.io/projected/4364f56d-f4a5-459c-9e28-1dd075d9a599-kube-api-access-wb95q\") pod \"keystone-operator-index-6wnsf\" (UID: \"4364f56d-f4a5-459c-9e28-1dd075d9a599\") " pod="openstack-operators/keystone-operator-index-6wnsf" Feb 02 15:30:58 crc kubenswrapper[4733]: I0202 15:30:58.751572 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wb95q\" (UniqueName: \"kubernetes.io/projected/4364f56d-f4a5-459c-9e28-1dd075d9a599-kube-api-access-wb95q\") pod \"keystone-operator-index-6wnsf\" (UID: \"4364f56d-f4a5-459c-9e28-1dd075d9a599\") " pod="openstack-operators/keystone-operator-index-6wnsf" Feb 02 15:30:58 crc kubenswrapper[4733]: I0202 15:30:58.779259 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wb95q\" (UniqueName: \"kubernetes.io/projected/4364f56d-f4a5-459c-9e28-1dd075d9a599-kube-api-access-wb95q\") pod \"keystone-operator-index-6wnsf\" (UID: \"4364f56d-f4a5-459c-9e28-1dd075d9a599\") " pod="openstack-operators/keystone-operator-index-6wnsf" Feb 02 15:30:58 crc kubenswrapper[4733]: I0202 15:30:58.852193 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-6wnsf" Feb 02 15:30:59 crc kubenswrapper[4733]: I0202 15:30:59.338253 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-6wnsf"] Feb 02 15:30:59 crc kubenswrapper[4733]: W0202 15:30:59.347413 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4364f56d_f4a5_459c_9e28_1dd075d9a599.slice/crio-6391588848a697f680f441b9ae3d3a1722f9f22a7ef57e205d6a5fd956e62b24 WatchSource:0}: Error finding container 6391588848a697f680f441b9ae3d3a1722f9f22a7ef57e205d6a5fd956e62b24: Status 404 returned error can't find the container with id 6391588848a697f680f441b9ae3d3a1722f9f22a7ef57e205d6a5fd956e62b24 Feb 02 15:31:00 crc kubenswrapper[4733]: I0202 15:31:00.167786 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-6wnsf" event={"ID":"4364f56d-f4a5-459c-9e28-1dd075d9a599","Type":"ContainerStarted","Data":"6391588848a697f680f441b9ae3d3a1722f9f22a7ef57e205d6a5fd956e62b24"} Feb 02 15:31:01 crc kubenswrapper[4733]: I0202 15:31:01.176150 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-6wnsf" event={"ID":"4364f56d-f4a5-459c-9e28-1dd075d9a599","Type":"ContainerStarted","Data":"97b8e5c8e218c080d4b76768bb029e8377bb04eefd64e3ad232159434ae27c63"} Feb 02 15:31:01 crc kubenswrapper[4733]: I0202 15:31:01.194523 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-index-6wnsf" podStartSLOduration=2.325764761 podStartE2EDuration="3.194504635s" podCreationTimestamp="2026-02-02 15:30:58 +0000 UTC" firstStartedPulling="2026-02-02 15:30:59.350631233 +0000 UTC m=+942.802092591" lastFinishedPulling="2026-02-02 15:31:00.219371107 +0000 UTC m=+943.670832465" observedRunningTime="2026-02-02 15:31:01.193840647 +0000 UTC m=+944.645302045" watchObservedRunningTime="2026-02-02 15:31:01.194504635 +0000 UTC m=+944.645965993" Feb 02 15:31:02 crc kubenswrapper[4733]: I0202 15:31:02.719432 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-index-6wnsf"] Feb 02 15:31:03 crc kubenswrapper[4733]: I0202 15:31:03.189244 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/keystone-operator-index-6wnsf" podUID="4364f56d-f4a5-459c-9e28-1dd075d9a599" containerName="registry-server" containerID="cri-o://97b8e5c8e218c080d4b76768bb029e8377bb04eefd64e3ad232159434ae27c63" gracePeriod=2 Feb 02 15:31:03 crc kubenswrapper[4733]: I0202 15:31:03.341679 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-index-l4jcf"] Feb 02 15:31:03 crc kubenswrapper[4733]: I0202 15:31:03.342691 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-l4jcf" Feb 02 15:31:03 crc kubenswrapper[4733]: I0202 15:31:03.352707 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-l4jcf"] Feb 02 15:31:03 crc kubenswrapper[4733]: I0202 15:31:03.424096 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dl76v\" (UniqueName: \"kubernetes.io/projected/d0147e51-446d-4a71-aaaa-e852258495d4-kube-api-access-dl76v\") pod \"keystone-operator-index-l4jcf\" (UID: \"d0147e51-446d-4a71-aaaa-e852258495d4\") " pod="openstack-operators/keystone-operator-index-l4jcf" Feb 02 15:31:03 crc kubenswrapper[4733]: I0202 15:31:03.525501 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dl76v\" (UniqueName: \"kubernetes.io/projected/d0147e51-446d-4a71-aaaa-e852258495d4-kube-api-access-dl76v\") pod \"keystone-operator-index-l4jcf\" (UID: \"d0147e51-446d-4a71-aaaa-e852258495d4\") " pod="openstack-operators/keystone-operator-index-l4jcf" Feb 02 15:31:03 crc kubenswrapper[4733]: I0202 15:31:03.548997 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dl76v\" (UniqueName: \"kubernetes.io/projected/d0147e51-446d-4a71-aaaa-e852258495d4-kube-api-access-dl76v\") pod \"keystone-operator-index-l4jcf\" (UID: \"d0147e51-446d-4a71-aaaa-e852258495d4\") " pod="openstack-operators/keystone-operator-index-l4jcf" Feb 02 15:31:03 crc kubenswrapper[4733]: I0202 15:31:03.683786 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-l4jcf" Feb 02 15:31:03 crc kubenswrapper[4733]: I0202 15:31:03.734212 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-6wnsf" Feb 02 15:31:03 crc kubenswrapper[4733]: I0202 15:31:03.829315 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wb95q\" (UniqueName: \"kubernetes.io/projected/4364f56d-f4a5-459c-9e28-1dd075d9a599-kube-api-access-wb95q\") pod \"4364f56d-f4a5-459c-9e28-1dd075d9a599\" (UID: \"4364f56d-f4a5-459c-9e28-1dd075d9a599\") " Feb 02 15:31:03 crc kubenswrapper[4733]: I0202 15:31:03.834209 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4364f56d-f4a5-459c-9e28-1dd075d9a599-kube-api-access-wb95q" (OuterVolumeSpecName: "kube-api-access-wb95q") pod "4364f56d-f4a5-459c-9e28-1dd075d9a599" (UID: "4364f56d-f4a5-459c-9e28-1dd075d9a599"). InnerVolumeSpecName "kube-api-access-wb95q". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:31:03 crc kubenswrapper[4733]: I0202 15:31:03.930874 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wb95q\" (UniqueName: \"kubernetes.io/projected/4364f56d-f4a5-459c-9e28-1dd075d9a599-kube-api-access-wb95q\") on node \"crc\" DevicePath \"\"" Feb 02 15:31:04 crc kubenswrapper[4733]: I0202 15:31:04.130051 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-l4jcf"] Feb 02 15:31:04 crc kubenswrapper[4733]: W0202 15:31:04.134716 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd0147e51_446d_4a71_aaaa_e852258495d4.slice/crio-5545c2b365737082b809c4c8d0bc0c975465df9799553b7895fda48062e0eecb WatchSource:0}: Error finding container 5545c2b365737082b809c4c8d0bc0c975465df9799553b7895fda48062e0eecb: Status 404 returned error can't find the container with id 5545c2b365737082b809c4c8d0bc0c975465df9799553b7895fda48062e0eecb Feb 02 15:31:04 crc kubenswrapper[4733]: I0202 15:31:04.198135 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-l4jcf" event={"ID":"d0147e51-446d-4a71-aaaa-e852258495d4","Type":"ContainerStarted","Data":"5545c2b365737082b809c4c8d0bc0c975465df9799553b7895fda48062e0eecb"} Feb 02 15:31:04 crc kubenswrapper[4733]: I0202 15:31:04.200940 4733 generic.go:334] "Generic (PLEG): container finished" podID="4364f56d-f4a5-459c-9e28-1dd075d9a599" containerID="97b8e5c8e218c080d4b76768bb029e8377bb04eefd64e3ad232159434ae27c63" exitCode=0 Feb 02 15:31:04 crc kubenswrapper[4733]: I0202 15:31:04.201455 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-6wnsf" event={"ID":"4364f56d-f4a5-459c-9e28-1dd075d9a599","Type":"ContainerDied","Data":"97b8e5c8e218c080d4b76768bb029e8377bb04eefd64e3ad232159434ae27c63"} Feb 02 15:31:04 crc kubenswrapper[4733]: I0202 15:31:04.204393 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-6wnsf" event={"ID":"4364f56d-f4a5-459c-9e28-1dd075d9a599","Type":"ContainerDied","Data":"6391588848a697f680f441b9ae3d3a1722f9f22a7ef57e205d6a5fd956e62b24"} Feb 02 15:31:04 crc kubenswrapper[4733]: I0202 15:31:04.204425 4733 scope.go:117] "RemoveContainer" containerID="97b8e5c8e218c080d4b76768bb029e8377bb04eefd64e3ad232159434ae27c63" Feb 02 15:31:04 crc kubenswrapper[4733]: I0202 15:31:04.201539 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-6wnsf" Feb 02 15:31:04 crc kubenswrapper[4733]: I0202 15:31:04.242869 4733 scope.go:117] "RemoveContainer" containerID="97b8e5c8e218c080d4b76768bb029e8377bb04eefd64e3ad232159434ae27c63" Feb 02 15:31:04 crc kubenswrapper[4733]: E0202 15:31:04.243311 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97b8e5c8e218c080d4b76768bb029e8377bb04eefd64e3ad232159434ae27c63\": container with ID starting with 97b8e5c8e218c080d4b76768bb029e8377bb04eefd64e3ad232159434ae27c63 not found: ID does not exist" containerID="97b8e5c8e218c080d4b76768bb029e8377bb04eefd64e3ad232159434ae27c63" Feb 02 15:31:04 crc kubenswrapper[4733]: I0202 15:31:04.243354 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97b8e5c8e218c080d4b76768bb029e8377bb04eefd64e3ad232159434ae27c63"} err="failed to get container status \"97b8e5c8e218c080d4b76768bb029e8377bb04eefd64e3ad232159434ae27c63\": rpc error: code = NotFound desc = could not find container \"97b8e5c8e218c080d4b76768bb029e8377bb04eefd64e3ad232159434ae27c63\": container with ID starting with 97b8e5c8e218c080d4b76768bb029e8377bb04eefd64e3ad232159434ae27c63 not found: ID does not exist" Feb 02 15:31:04 crc kubenswrapper[4733]: I0202 15:31:04.250721 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-index-6wnsf"] Feb 02 15:31:04 crc kubenswrapper[4733]: I0202 15:31:04.257201 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/keystone-operator-index-6wnsf"] Feb 02 15:31:05 crc kubenswrapper[4733]: I0202 15:31:05.211910 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-l4jcf" event={"ID":"d0147e51-446d-4a71-aaaa-e852258495d4","Type":"ContainerStarted","Data":"e8fd5017367c208f4e3389d15fb93e2782229ba4ab3d82f8fefe713594ebcad1"} Feb 02 15:31:05 crc kubenswrapper[4733]: I0202 15:31:05.238278 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-index-l4jcf" podStartSLOduration=1.831681615 podStartE2EDuration="2.238244544s" podCreationTimestamp="2026-02-02 15:31:03 +0000 UTC" firstStartedPulling="2026-02-02 15:31:04.138730077 +0000 UTC m=+947.590191435" lastFinishedPulling="2026-02-02 15:31:04.545293006 +0000 UTC m=+947.996754364" observedRunningTime="2026-02-02 15:31:05.231339151 +0000 UTC m=+948.682800519" watchObservedRunningTime="2026-02-02 15:31:05.238244544 +0000 UTC m=+948.689705942" Feb 02 15:31:05 crc kubenswrapper[4733]: I0202 15:31:05.266773 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4364f56d-f4a5-459c-9e28-1dd075d9a599" path="/var/lib/kubelet/pods/4364f56d-f4a5-459c-9e28-1dd075d9a599/volumes" Feb 02 15:31:07 crc kubenswrapper[4733]: I0202 15:31:07.310295 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["manila-kuttl-tests/rabbitmq-server-0"] Feb 02 15:31:07 crc kubenswrapper[4733]: E0202 15:31:07.310613 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4364f56d-f4a5-459c-9e28-1dd075d9a599" containerName="registry-server" Feb 02 15:31:07 crc kubenswrapper[4733]: I0202 15:31:07.310631 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="4364f56d-f4a5-459c-9e28-1dd075d9a599" containerName="registry-server" Feb 02 15:31:07 crc kubenswrapper[4733]: I0202 15:31:07.310863 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="4364f56d-f4a5-459c-9e28-1dd075d9a599" containerName="registry-server" Feb 02 15:31:07 crc kubenswrapper[4733]: I0202 15:31:07.311852 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/rabbitmq-server-0" Feb 02 15:31:07 crc kubenswrapper[4733]: I0202 15:31:07.314497 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"rabbitmq-default-user" Feb 02 15:31:07 crc kubenswrapper[4733]: I0202 15:31:07.314937 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"manila-kuttl-tests"/"rabbitmq-plugins-conf" Feb 02 15:31:07 crc kubenswrapper[4733]: I0202 15:31:07.315196 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"rabbitmq-server-dockercfg-v6ck9" Feb 02 15:31:07 crc kubenswrapper[4733]: I0202 15:31:07.316445 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"manila-kuttl-tests"/"rabbitmq-server-conf" Feb 02 15:31:07 crc kubenswrapper[4733]: I0202 15:31:07.319550 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"rabbitmq-erlang-cookie" Feb 02 15:31:07 crc kubenswrapper[4733]: I0202 15:31:07.330418 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/rabbitmq-server-0"] Feb 02 15:31:07 crc kubenswrapper[4733]: I0202 15:31:07.387682 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f45e90dd-ab9d-4c27-8449-98cf52f48c87-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f45e90dd-ab9d-4c27-8449-98cf52f48c87\") " pod="manila-kuttl-tests/rabbitmq-server-0" Feb 02 15:31:07 crc kubenswrapper[4733]: I0202 15:31:07.388004 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f45e90dd-ab9d-4c27-8449-98cf52f48c87-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f45e90dd-ab9d-4c27-8449-98cf52f48c87\") " pod="manila-kuttl-tests/rabbitmq-server-0" Feb 02 15:31:07 crc kubenswrapper[4733]: I0202 15:31:07.388124 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f45e90dd-ab9d-4c27-8449-98cf52f48c87-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f45e90dd-ab9d-4c27-8449-98cf52f48c87\") " pod="manila-kuttl-tests/rabbitmq-server-0" Feb 02 15:31:07 crc kubenswrapper[4733]: I0202 15:31:07.388280 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f45e90dd-ab9d-4c27-8449-98cf52f48c87-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f45e90dd-ab9d-4c27-8449-98cf52f48c87\") " pod="manila-kuttl-tests/rabbitmq-server-0" Feb 02 15:31:07 crc kubenswrapper[4733]: I0202 15:31:07.388411 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfh6r\" (UniqueName: \"kubernetes.io/projected/f45e90dd-ab9d-4c27-8449-98cf52f48c87-kube-api-access-mfh6r\") pod \"rabbitmq-server-0\" (UID: \"f45e90dd-ab9d-4c27-8449-98cf52f48c87\") " pod="manila-kuttl-tests/rabbitmq-server-0" Feb 02 15:31:07 crc kubenswrapper[4733]: I0202 15:31:07.388571 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-cee723a2-19bb-4ddf-b89c-19eee3872720\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cee723a2-19bb-4ddf-b89c-19eee3872720\") pod \"rabbitmq-server-0\" (UID: \"f45e90dd-ab9d-4c27-8449-98cf52f48c87\") " pod="manila-kuttl-tests/rabbitmq-server-0" Feb 02 15:31:07 crc kubenswrapper[4733]: I0202 15:31:07.388726 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f45e90dd-ab9d-4c27-8449-98cf52f48c87-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f45e90dd-ab9d-4c27-8449-98cf52f48c87\") " pod="manila-kuttl-tests/rabbitmq-server-0" Feb 02 15:31:07 crc kubenswrapper[4733]: I0202 15:31:07.388859 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f45e90dd-ab9d-4c27-8449-98cf52f48c87-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f45e90dd-ab9d-4c27-8449-98cf52f48c87\") " pod="manila-kuttl-tests/rabbitmq-server-0" Feb 02 15:31:07 crc kubenswrapper[4733]: I0202 15:31:07.490055 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f45e90dd-ab9d-4c27-8449-98cf52f48c87-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f45e90dd-ab9d-4c27-8449-98cf52f48c87\") " pod="manila-kuttl-tests/rabbitmq-server-0" Feb 02 15:31:07 crc kubenswrapper[4733]: I0202 15:31:07.490210 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f45e90dd-ab9d-4c27-8449-98cf52f48c87-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f45e90dd-ab9d-4c27-8449-98cf52f48c87\") " pod="manila-kuttl-tests/rabbitmq-server-0" Feb 02 15:31:07 crc kubenswrapper[4733]: I0202 15:31:07.490250 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f45e90dd-ab9d-4c27-8449-98cf52f48c87-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f45e90dd-ab9d-4c27-8449-98cf52f48c87\") " pod="manila-kuttl-tests/rabbitmq-server-0" Feb 02 15:31:07 crc kubenswrapper[4733]: I0202 15:31:07.490316 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f45e90dd-ab9d-4c27-8449-98cf52f48c87-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f45e90dd-ab9d-4c27-8449-98cf52f48c87\") " pod="manila-kuttl-tests/rabbitmq-server-0" Feb 02 15:31:07 crc kubenswrapper[4733]: I0202 15:31:07.490353 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfh6r\" (UniqueName: \"kubernetes.io/projected/f45e90dd-ab9d-4c27-8449-98cf52f48c87-kube-api-access-mfh6r\") pod \"rabbitmq-server-0\" (UID: \"f45e90dd-ab9d-4c27-8449-98cf52f48c87\") " pod="manila-kuttl-tests/rabbitmq-server-0" Feb 02 15:31:07 crc kubenswrapper[4733]: I0202 15:31:07.490389 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-cee723a2-19bb-4ddf-b89c-19eee3872720\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cee723a2-19bb-4ddf-b89c-19eee3872720\") pod \"rabbitmq-server-0\" (UID: \"f45e90dd-ab9d-4c27-8449-98cf52f48c87\") " pod="manila-kuttl-tests/rabbitmq-server-0" Feb 02 15:31:07 crc kubenswrapper[4733]: I0202 15:31:07.490443 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f45e90dd-ab9d-4c27-8449-98cf52f48c87-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f45e90dd-ab9d-4c27-8449-98cf52f48c87\") " pod="manila-kuttl-tests/rabbitmq-server-0" Feb 02 15:31:07 crc kubenswrapper[4733]: I0202 15:31:07.490475 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f45e90dd-ab9d-4c27-8449-98cf52f48c87-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f45e90dd-ab9d-4c27-8449-98cf52f48c87\") " pod="manila-kuttl-tests/rabbitmq-server-0" Feb 02 15:31:07 crc kubenswrapper[4733]: I0202 15:31:07.490994 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f45e90dd-ab9d-4c27-8449-98cf52f48c87-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f45e90dd-ab9d-4c27-8449-98cf52f48c87\") " pod="manila-kuttl-tests/rabbitmq-server-0" Feb 02 15:31:07 crc kubenswrapper[4733]: I0202 15:31:07.491205 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f45e90dd-ab9d-4c27-8449-98cf52f48c87-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f45e90dd-ab9d-4c27-8449-98cf52f48c87\") " pod="manila-kuttl-tests/rabbitmq-server-0" Feb 02 15:31:07 crc kubenswrapper[4733]: I0202 15:31:07.492378 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f45e90dd-ab9d-4c27-8449-98cf52f48c87-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f45e90dd-ab9d-4c27-8449-98cf52f48c87\") " pod="manila-kuttl-tests/rabbitmq-server-0" Feb 02 15:31:07 crc kubenswrapper[4733]: I0202 15:31:07.497405 4733 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 02 15:31:07 crc kubenswrapper[4733]: I0202 15:31:07.497477 4733 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-cee723a2-19bb-4ddf-b89c-19eee3872720\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cee723a2-19bb-4ddf-b89c-19eee3872720\") pod \"rabbitmq-server-0\" (UID: \"f45e90dd-ab9d-4c27-8449-98cf52f48c87\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/8e92fe1d3b14d1bae2a69a35ef9d761be8acdf21f7b2fa1fce7917b4c845071c/globalmount\"" pod="manila-kuttl-tests/rabbitmq-server-0" Feb 02 15:31:07 crc kubenswrapper[4733]: I0202 15:31:07.497601 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f45e90dd-ab9d-4c27-8449-98cf52f48c87-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f45e90dd-ab9d-4c27-8449-98cf52f48c87\") " pod="manila-kuttl-tests/rabbitmq-server-0" Feb 02 15:31:07 crc kubenswrapper[4733]: I0202 15:31:07.497652 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f45e90dd-ab9d-4c27-8449-98cf52f48c87-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f45e90dd-ab9d-4c27-8449-98cf52f48c87\") " pod="manila-kuttl-tests/rabbitmq-server-0" Feb 02 15:31:07 crc kubenswrapper[4733]: I0202 15:31:07.498703 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f45e90dd-ab9d-4c27-8449-98cf52f48c87-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f45e90dd-ab9d-4c27-8449-98cf52f48c87\") " pod="manila-kuttl-tests/rabbitmq-server-0" Feb 02 15:31:07 crc kubenswrapper[4733]: I0202 15:31:07.511303 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfh6r\" (UniqueName: \"kubernetes.io/projected/f45e90dd-ab9d-4c27-8449-98cf52f48c87-kube-api-access-mfh6r\") pod \"rabbitmq-server-0\" (UID: \"f45e90dd-ab9d-4c27-8449-98cf52f48c87\") " pod="manila-kuttl-tests/rabbitmq-server-0" Feb 02 15:31:07 crc kubenswrapper[4733]: I0202 15:31:07.532412 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-cee723a2-19bb-4ddf-b89c-19eee3872720\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cee723a2-19bb-4ddf-b89c-19eee3872720\") pod \"rabbitmq-server-0\" (UID: \"f45e90dd-ab9d-4c27-8449-98cf52f48c87\") " pod="manila-kuttl-tests/rabbitmq-server-0" Feb 02 15:31:07 crc kubenswrapper[4733]: I0202 15:31:07.648936 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/rabbitmq-server-0" Feb 02 15:31:08 crc kubenswrapper[4733]: I0202 15:31:08.127070 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/rabbitmq-server-0"] Feb 02 15:31:08 crc kubenswrapper[4733]: I0202 15:31:08.236610 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/rabbitmq-server-0" event={"ID":"f45e90dd-ab9d-4c27-8449-98cf52f48c87","Type":"ContainerStarted","Data":"ca0247762a0b2747313840f74a70bbe642746790e8a786c7dbfdc7b707a1a2a7"} Feb 02 15:31:13 crc kubenswrapper[4733]: I0202 15:31:13.684528 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/keystone-operator-index-l4jcf" Feb 02 15:31:13 crc kubenswrapper[4733]: I0202 15:31:13.685117 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-index-l4jcf" Feb 02 15:31:13 crc kubenswrapper[4733]: I0202 15:31:13.718122 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/keystone-operator-index-l4jcf" Feb 02 15:31:14 crc kubenswrapper[4733]: I0202 15:31:14.339954 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-index-l4jcf" Feb 02 15:31:15 crc kubenswrapper[4733]: I0202 15:31:15.787687 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqw8xf"] Feb 02 15:31:15 crc kubenswrapper[4733]: I0202 15:31:15.790064 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqw8xf" Feb 02 15:31:15 crc kubenswrapper[4733]: I0202 15:31:15.792505 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-2mtp8" Feb 02 15:31:15 crc kubenswrapper[4733]: I0202 15:31:15.802540 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqw8xf"] Feb 02 15:31:15 crc kubenswrapper[4733]: I0202 15:31:15.923689 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2fbae2a6-73a0-496e-900c-6d0d4c704994-util\") pod \"b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqw8xf\" (UID: \"2fbae2a6-73a0-496e-900c-6d0d4c704994\") " pod="openstack-operators/b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqw8xf" Feb 02 15:31:15 crc kubenswrapper[4733]: I0202 15:31:15.924210 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2fbae2a6-73a0-496e-900c-6d0d4c704994-bundle\") pod \"b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqw8xf\" (UID: \"2fbae2a6-73a0-496e-900c-6d0d4c704994\") " pod="openstack-operators/b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqw8xf" Feb 02 15:31:15 crc kubenswrapper[4733]: I0202 15:31:15.924570 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqqxk\" (UniqueName: \"kubernetes.io/projected/2fbae2a6-73a0-496e-900c-6d0d4c704994-kube-api-access-wqqxk\") pod \"b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqw8xf\" (UID: \"2fbae2a6-73a0-496e-900c-6d0d4c704994\") " pod="openstack-operators/b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqw8xf" Feb 02 15:31:16 crc kubenswrapper[4733]: I0202 15:31:16.025725 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wqqxk\" (UniqueName: \"kubernetes.io/projected/2fbae2a6-73a0-496e-900c-6d0d4c704994-kube-api-access-wqqxk\") pod \"b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqw8xf\" (UID: \"2fbae2a6-73a0-496e-900c-6d0d4c704994\") " pod="openstack-operators/b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqw8xf" Feb 02 15:31:16 crc kubenswrapper[4733]: I0202 15:31:16.025774 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2fbae2a6-73a0-496e-900c-6d0d4c704994-util\") pod \"b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqw8xf\" (UID: \"2fbae2a6-73a0-496e-900c-6d0d4c704994\") " pod="openstack-operators/b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqw8xf" Feb 02 15:31:16 crc kubenswrapper[4733]: I0202 15:31:16.025828 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2fbae2a6-73a0-496e-900c-6d0d4c704994-bundle\") pod \"b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqw8xf\" (UID: \"2fbae2a6-73a0-496e-900c-6d0d4c704994\") " pod="openstack-operators/b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqw8xf" Feb 02 15:31:16 crc kubenswrapper[4733]: I0202 15:31:16.026343 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2fbae2a6-73a0-496e-900c-6d0d4c704994-util\") pod \"b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqw8xf\" (UID: \"2fbae2a6-73a0-496e-900c-6d0d4c704994\") " pod="openstack-operators/b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqw8xf" Feb 02 15:31:16 crc kubenswrapper[4733]: I0202 15:31:16.026419 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2fbae2a6-73a0-496e-900c-6d0d4c704994-bundle\") pod \"b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqw8xf\" (UID: \"2fbae2a6-73a0-496e-900c-6d0d4c704994\") " pod="openstack-operators/b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqw8xf" Feb 02 15:31:16 crc kubenswrapper[4733]: I0202 15:31:16.045646 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wqqxk\" (UniqueName: \"kubernetes.io/projected/2fbae2a6-73a0-496e-900c-6d0d4c704994-kube-api-access-wqqxk\") pod \"b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqw8xf\" (UID: \"2fbae2a6-73a0-496e-900c-6d0d4c704994\") " pod="openstack-operators/b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqw8xf" Feb 02 15:31:16 crc kubenswrapper[4733]: I0202 15:31:16.109385 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqw8xf" Feb 02 15:31:16 crc kubenswrapper[4733]: I0202 15:31:16.304971 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/rabbitmq-server-0" event={"ID":"f45e90dd-ab9d-4c27-8449-98cf52f48c87","Type":"ContainerStarted","Data":"9a9ed87542bec5386676f7a6416db7de1ec8fc04680e6dbf0f22482fb372172c"} Feb 02 15:31:16 crc kubenswrapper[4733]: I0202 15:31:16.557093 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqw8xf"] Feb 02 15:31:16 crc kubenswrapper[4733]: W0202 15:31:16.557331 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2fbae2a6_73a0_496e_900c_6d0d4c704994.slice/crio-94a2b7ad4b41fc198e95ff169f460914cb873962cb527d0d8c7d2cca8f7d3933 WatchSource:0}: Error finding container 94a2b7ad4b41fc198e95ff169f460914cb873962cb527d0d8c7d2cca8f7d3933: Status 404 returned error can't find the container with id 94a2b7ad4b41fc198e95ff169f460914cb873962cb527d0d8c7d2cca8f7d3933 Feb 02 15:31:17 crc kubenswrapper[4733]: I0202 15:31:17.317528 4733 generic.go:334] "Generic (PLEG): container finished" podID="2fbae2a6-73a0-496e-900c-6d0d4c704994" containerID="50bf7cb2672133cbe14cca992caced641278856ab52e9e82b6877bad6ca43b78" exitCode=0 Feb 02 15:31:17 crc kubenswrapper[4733]: I0202 15:31:17.318474 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqw8xf" event={"ID":"2fbae2a6-73a0-496e-900c-6d0d4c704994","Type":"ContainerDied","Data":"50bf7cb2672133cbe14cca992caced641278856ab52e9e82b6877bad6ca43b78"} Feb 02 15:31:17 crc kubenswrapper[4733]: I0202 15:31:17.318508 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqw8xf" event={"ID":"2fbae2a6-73a0-496e-900c-6d0d4c704994","Type":"ContainerStarted","Data":"94a2b7ad4b41fc198e95ff169f460914cb873962cb527d0d8c7d2cca8f7d3933"} Feb 02 15:31:19 crc kubenswrapper[4733]: I0202 15:31:19.336493 4733 generic.go:334] "Generic (PLEG): container finished" podID="2fbae2a6-73a0-496e-900c-6d0d4c704994" containerID="5a1d2d8c4e19b20ac57cb8b31c117bc5dfc5e552e0082061c2b690e63e76f1c8" exitCode=0 Feb 02 15:31:19 crc kubenswrapper[4733]: I0202 15:31:19.336615 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqw8xf" event={"ID":"2fbae2a6-73a0-496e-900c-6d0d4c704994","Type":"ContainerDied","Data":"5a1d2d8c4e19b20ac57cb8b31c117bc5dfc5e552e0082061c2b690e63e76f1c8"} Feb 02 15:31:20 crc kubenswrapper[4733]: I0202 15:31:20.367528 4733 generic.go:334] "Generic (PLEG): container finished" podID="2fbae2a6-73a0-496e-900c-6d0d4c704994" containerID="b7e3158461aaa2f50e0cf874860750cd5ffae4b2e58750199b1253c60286b0fc" exitCode=0 Feb 02 15:31:20 crc kubenswrapper[4733]: I0202 15:31:20.367987 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqw8xf" event={"ID":"2fbae2a6-73a0-496e-900c-6d0d4c704994","Type":"ContainerDied","Data":"b7e3158461aaa2f50e0cf874860750cd5ffae4b2e58750199b1253c60286b0fc"} Feb 02 15:31:21 crc kubenswrapper[4733]: I0202 15:31:21.689732 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqw8xf" Feb 02 15:31:21 crc kubenswrapper[4733]: I0202 15:31:21.810938 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wqqxk\" (UniqueName: \"kubernetes.io/projected/2fbae2a6-73a0-496e-900c-6d0d4c704994-kube-api-access-wqqxk\") pod \"2fbae2a6-73a0-496e-900c-6d0d4c704994\" (UID: \"2fbae2a6-73a0-496e-900c-6d0d4c704994\") " Feb 02 15:31:21 crc kubenswrapper[4733]: I0202 15:31:21.811211 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2fbae2a6-73a0-496e-900c-6d0d4c704994-util\") pod \"2fbae2a6-73a0-496e-900c-6d0d4c704994\" (UID: \"2fbae2a6-73a0-496e-900c-6d0d4c704994\") " Feb 02 15:31:21 crc kubenswrapper[4733]: I0202 15:31:21.811277 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2fbae2a6-73a0-496e-900c-6d0d4c704994-bundle\") pod \"2fbae2a6-73a0-496e-900c-6d0d4c704994\" (UID: \"2fbae2a6-73a0-496e-900c-6d0d4c704994\") " Feb 02 15:31:21 crc kubenswrapper[4733]: I0202 15:31:21.813467 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2fbae2a6-73a0-496e-900c-6d0d4c704994-bundle" (OuterVolumeSpecName: "bundle") pod "2fbae2a6-73a0-496e-900c-6d0d4c704994" (UID: "2fbae2a6-73a0-496e-900c-6d0d4c704994"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:31:21 crc kubenswrapper[4733]: I0202 15:31:21.818190 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fbae2a6-73a0-496e-900c-6d0d4c704994-kube-api-access-wqqxk" (OuterVolumeSpecName: "kube-api-access-wqqxk") pod "2fbae2a6-73a0-496e-900c-6d0d4c704994" (UID: "2fbae2a6-73a0-496e-900c-6d0d4c704994"). InnerVolumeSpecName "kube-api-access-wqqxk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:31:21 crc kubenswrapper[4733]: I0202 15:31:21.841794 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2fbae2a6-73a0-496e-900c-6d0d4c704994-util" (OuterVolumeSpecName: "util") pod "2fbae2a6-73a0-496e-900c-6d0d4c704994" (UID: "2fbae2a6-73a0-496e-900c-6d0d4c704994"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:31:21 crc kubenswrapper[4733]: I0202 15:31:21.912953 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wqqxk\" (UniqueName: \"kubernetes.io/projected/2fbae2a6-73a0-496e-900c-6d0d4c704994-kube-api-access-wqqxk\") on node \"crc\" DevicePath \"\"" Feb 02 15:31:21 crc kubenswrapper[4733]: I0202 15:31:21.913009 4733 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2fbae2a6-73a0-496e-900c-6d0d4c704994-util\") on node \"crc\" DevicePath \"\"" Feb 02 15:31:21 crc kubenswrapper[4733]: I0202 15:31:21.913034 4733 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2fbae2a6-73a0-496e-900c-6d0d4c704994-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 15:31:22 crc kubenswrapper[4733]: I0202 15:31:22.387645 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqw8xf" event={"ID":"2fbae2a6-73a0-496e-900c-6d0d4c704994","Type":"ContainerDied","Data":"94a2b7ad4b41fc198e95ff169f460914cb873962cb527d0d8c7d2cca8f7d3933"} Feb 02 15:31:22 crc kubenswrapper[4733]: I0202 15:31:22.387735 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="94a2b7ad4b41fc198e95ff169f460914cb873962cb527d0d8c7d2cca8f7d3933" Feb 02 15:31:22 crc kubenswrapper[4733]: I0202 15:31:22.387785 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqw8xf" Feb 02 15:31:41 crc kubenswrapper[4733]: I0202 15:31:41.597549 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-86bf5ddb6d-lx2j7"] Feb 02 15:31:41 crc kubenswrapper[4733]: E0202 15:31:41.599339 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fbae2a6-73a0-496e-900c-6d0d4c704994" containerName="pull" Feb 02 15:31:41 crc kubenswrapper[4733]: I0202 15:31:41.599431 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fbae2a6-73a0-496e-900c-6d0d4c704994" containerName="pull" Feb 02 15:31:41 crc kubenswrapper[4733]: E0202 15:31:41.599513 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fbae2a6-73a0-496e-900c-6d0d4c704994" containerName="util" Feb 02 15:31:41 crc kubenswrapper[4733]: I0202 15:31:41.599591 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fbae2a6-73a0-496e-900c-6d0d4c704994" containerName="util" Feb 02 15:31:41 crc kubenswrapper[4733]: E0202 15:31:41.599671 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fbae2a6-73a0-496e-900c-6d0d4c704994" containerName="extract" Feb 02 15:31:41 crc kubenswrapper[4733]: I0202 15:31:41.599743 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fbae2a6-73a0-496e-900c-6d0d4c704994" containerName="extract" Feb 02 15:31:41 crc kubenswrapper[4733]: I0202 15:31:41.599972 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fbae2a6-73a0-496e-900c-6d0d4c704994" containerName="extract" Feb 02 15:31:41 crc kubenswrapper[4733]: I0202 15:31:41.600582 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-86bf5ddb6d-lx2j7" Feb 02 15:31:41 crc kubenswrapper[4733]: I0202 15:31:41.604580 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-j6msk" Feb 02 15:31:41 crc kubenswrapper[4733]: I0202 15:31:41.604971 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-service-cert" Feb 02 15:31:41 crc kubenswrapper[4733]: I0202 15:31:41.623309 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-86bf5ddb6d-lx2j7"] Feb 02 15:31:41 crc kubenswrapper[4733]: I0202 15:31:41.786609 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1aaf01fd-0114-41a6-af39-bd1a10deabe9-webhook-cert\") pod \"keystone-operator-controller-manager-86bf5ddb6d-lx2j7\" (UID: \"1aaf01fd-0114-41a6-af39-bd1a10deabe9\") " pod="openstack-operators/keystone-operator-controller-manager-86bf5ddb6d-lx2j7" Feb 02 15:31:41 crc kubenswrapper[4733]: I0202 15:31:41.786799 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1aaf01fd-0114-41a6-af39-bd1a10deabe9-apiservice-cert\") pod \"keystone-operator-controller-manager-86bf5ddb6d-lx2j7\" (UID: \"1aaf01fd-0114-41a6-af39-bd1a10deabe9\") " pod="openstack-operators/keystone-operator-controller-manager-86bf5ddb6d-lx2j7" Feb 02 15:31:41 crc kubenswrapper[4733]: I0202 15:31:41.786885 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4qzm\" (UniqueName: \"kubernetes.io/projected/1aaf01fd-0114-41a6-af39-bd1a10deabe9-kube-api-access-d4qzm\") pod \"keystone-operator-controller-manager-86bf5ddb6d-lx2j7\" (UID: \"1aaf01fd-0114-41a6-af39-bd1a10deabe9\") " pod="openstack-operators/keystone-operator-controller-manager-86bf5ddb6d-lx2j7" Feb 02 15:31:41 crc kubenswrapper[4733]: I0202 15:31:41.888514 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1aaf01fd-0114-41a6-af39-bd1a10deabe9-apiservice-cert\") pod \"keystone-operator-controller-manager-86bf5ddb6d-lx2j7\" (UID: \"1aaf01fd-0114-41a6-af39-bd1a10deabe9\") " pod="openstack-operators/keystone-operator-controller-manager-86bf5ddb6d-lx2j7" Feb 02 15:31:41 crc kubenswrapper[4733]: I0202 15:31:41.888649 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4qzm\" (UniqueName: \"kubernetes.io/projected/1aaf01fd-0114-41a6-af39-bd1a10deabe9-kube-api-access-d4qzm\") pod \"keystone-operator-controller-manager-86bf5ddb6d-lx2j7\" (UID: \"1aaf01fd-0114-41a6-af39-bd1a10deabe9\") " pod="openstack-operators/keystone-operator-controller-manager-86bf5ddb6d-lx2j7" Feb 02 15:31:41 crc kubenswrapper[4733]: I0202 15:31:41.888707 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1aaf01fd-0114-41a6-af39-bd1a10deabe9-webhook-cert\") pod \"keystone-operator-controller-manager-86bf5ddb6d-lx2j7\" (UID: \"1aaf01fd-0114-41a6-af39-bd1a10deabe9\") " pod="openstack-operators/keystone-operator-controller-manager-86bf5ddb6d-lx2j7" Feb 02 15:31:41 crc kubenswrapper[4733]: I0202 15:31:41.894654 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1aaf01fd-0114-41a6-af39-bd1a10deabe9-webhook-cert\") pod \"keystone-operator-controller-manager-86bf5ddb6d-lx2j7\" (UID: \"1aaf01fd-0114-41a6-af39-bd1a10deabe9\") " pod="openstack-operators/keystone-operator-controller-manager-86bf5ddb6d-lx2j7" Feb 02 15:31:41 crc kubenswrapper[4733]: I0202 15:31:41.898798 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1aaf01fd-0114-41a6-af39-bd1a10deabe9-apiservice-cert\") pod \"keystone-operator-controller-manager-86bf5ddb6d-lx2j7\" (UID: \"1aaf01fd-0114-41a6-af39-bd1a10deabe9\") " pod="openstack-operators/keystone-operator-controller-manager-86bf5ddb6d-lx2j7" Feb 02 15:31:41 crc kubenswrapper[4733]: I0202 15:31:41.910651 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4qzm\" (UniqueName: \"kubernetes.io/projected/1aaf01fd-0114-41a6-af39-bd1a10deabe9-kube-api-access-d4qzm\") pod \"keystone-operator-controller-manager-86bf5ddb6d-lx2j7\" (UID: \"1aaf01fd-0114-41a6-af39-bd1a10deabe9\") " pod="openstack-operators/keystone-operator-controller-manager-86bf5ddb6d-lx2j7" Feb 02 15:31:41 crc kubenswrapper[4733]: I0202 15:31:41.922190 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-86bf5ddb6d-lx2j7" Feb 02 15:31:42 crc kubenswrapper[4733]: I0202 15:31:42.371079 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-86bf5ddb6d-lx2j7"] Feb 02 15:31:42 crc kubenswrapper[4733]: I0202 15:31:42.526060 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-86bf5ddb6d-lx2j7" event={"ID":"1aaf01fd-0114-41a6-af39-bd1a10deabe9","Type":"ContainerStarted","Data":"994566ca492a57f666784fc02b02c3ef6f5012b842546007a5985ae6268d8324"} Feb 02 15:31:47 crc kubenswrapper[4733]: I0202 15:31:47.556921 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-86bf5ddb6d-lx2j7" event={"ID":"1aaf01fd-0114-41a6-af39-bd1a10deabe9","Type":"ContainerStarted","Data":"a81e99b0593d8b4fc3e8efc759cf74ef72b91cef1303d352fdc881f4c7aa0104"} Feb 02 15:31:47 crc kubenswrapper[4733]: I0202 15:31:47.557379 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-86bf5ddb6d-lx2j7" Feb 02 15:31:47 crc kubenswrapper[4733]: I0202 15:31:47.577594 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-86bf5ddb6d-lx2j7" podStartSLOduration=2.28993648 podStartE2EDuration="6.577572823s" podCreationTimestamp="2026-02-02 15:31:41 +0000 UTC" firstStartedPulling="2026-02-02 15:31:42.373055038 +0000 UTC m=+985.824516406" lastFinishedPulling="2026-02-02 15:31:46.660691391 +0000 UTC m=+990.112152749" observedRunningTime="2026-02-02 15:31:47.574622584 +0000 UTC m=+991.026083942" watchObservedRunningTime="2026-02-02 15:31:47.577572823 +0000 UTC m=+991.029034181" Feb 02 15:31:48 crc kubenswrapper[4733]: I0202 15:31:48.568340 4733 generic.go:334] "Generic (PLEG): container finished" podID="f45e90dd-ab9d-4c27-8449-98cf52f48c87" containerID="9a9ed87542bec5386676f7a6416db7de1ec8fc04680e6dbf0f22482fb372172c" exitCode=0 Feb 02 15:31:48 crc kubenswrapper[4733]: I0202 15:31:48.568375 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/rabbitmq-server-0" event={"ID":"f45e90dd-ab9d-4c27-8449-98cf52f48c87","Type":"ContainerDied","Data":"9a9ed87542bec5386676f7a6416db7de1ec8fc04680e6dbf0f22482fb372172c"} Feb 02 15:31:49 crc kubenswrapper[4733]: I0202 15:31:49.579021 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/rabbitmq-server-0" event={"ID":"f45e90dd-ab9d-4c27-8449-98cf52f48c87","Type":"ContainerStarted","Data":"40d292ee4fa8872456f484269b6ff9ad31d0092c116a9313314a8f34ed13addd"} Feb 02 15:31:49 crc kubenswrapper[4733]: I0202 15:31:49.579518 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="manila-kuttl-tests/rabbitmq-server-0" Feb 02 15:31:49 crc kubenswrapper[4733]: I0202 15:31:49.617125 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="manila-kuttl-tests/rabbitmq-server-0" podStartSLOduration=37.439509059 podStartE2EDuration="43.617104686s" podCreationTimestamp="2026-02-02 15:31:06 +0000 UTC" firstStartedPulling="2026-02-02 15:31:08.139733943 +0000 UTC m=+951.591195311" lastFinishedPulling="2026-02-02 15:31:14.31732958 +0000 UTC m=+957.768790938" observedRunningTime="2026-02-02 15:31:49.612587326 +0000 UTC m=+993.064048704" watchObservedRunningTime="2026-02-02 15:31:49.617104686 +0000 UTC m=+993.068566044" Feb 02 15:31:51 crc kubenswrapper[4733]: I0202 15:31:51.927666 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-86bf5ddb6d-lx2j7" Feb 02 15:31:56 crc kubenswrapper[4733]: I0202 15:31:56.361328 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["manila-kuttl-tests/keystone-db-create-s5qkh"] Feb 02 15:31:56 crc kubenswrapper[4733]: I0202 15:31:56.363686 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/keystone-db-create-s5qkh" Feb 02 15:31:56 crc kubenswrapper[4733]: I0202 15:31:56.369050 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["manila-kuttl-tests/keystone-c686-account-create-update-msktt"] Feb 02 15:31:56 crc kubenswrapper[4733]: I0202 15:31:56.370225 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/keystone-c686-account-create-update-msktt" Feb 02 15:31:56 crc kubenswrapper[4733]: I0202 15:31:56.374143 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/keystone-db-create-s5qkh"] Feb 02 15:31:56 crc kubenswrapper[4733]: I0202 15:31:56.376324 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"keystone-db-secret" Feb 02 15:31:56 crc kubenswrapper[4733]: I0202 15:31:56.387087 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/keystone-c686-account-create-update-msktt"] Feb 02 15:31:56 crc kubenswrapper[4733]: I0202 15:31:56.499015 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/353d8d01-e2c8-40ec-80e4-034ab7786569-operator-scripts\") pod \"keystone-c686-account-create-update-msktt\" (UID: \"353d8d01-e2c8-40ec-80e4-034ab7786569\") " pod="manila-kuttl-tests/keystone-c686-account-create-update-msktt" Feb 02 15:31:56 crc kubenswrapper[4733]: I0202 15:31:56.499222 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/316f3a85-5cf5-44c5-b8f5-fbc6d7cbff3d-operator-scripts\") pod \"keystone-db-create-s5qkh\" (UID: \"316f3a85-5cf5-44c5-b8f5-fbc6d7cbff3d\") " pod="manila-kuttl-tests/keystone-db-create-s5qkh" Feb 02 15:31:56 crc kubenswrapper[4733]: I0202 15:31:56.499624 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8g7l4\" (UniqueName: \"kubernetes.io/projected/316f3a85-5cf5-44c5-b8f5-fbc6d7cbff3d-kube-api-access-8g7l4\") pod \"keystone-db-create-s5qkh\" (UID: \"316f3a85-5cf5-44c5-b8f5-fbc6d7cbff3d\") " pod="manila-kuttl-tests/keystone-db-create-s5qkh" Feb 02 15:31:56 crc kubenswrapper[4733]: I0202 15:31:56.499704 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4jg5\" (UniqueName: \"kubernetes.io/projected/353d8d01-e2c8-40ec-80e4-034ab7786569-kube-api-access-f4jg5\") pod \"keystone-c686-account-create-update-msktt\" (UID: \"353d8d01-e2c8-40ec-80e4-034ab7786569\") " pod="manila-kuttl-tests/keystone-c686-account-create-update-msktt" Feb 02 15:31:56 crc kubenswrapper[4733]: I0202 15:31:56.601616 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/316f3a85-5cf5-44c5-b8f5-fbc6d7cbff3d-operator-scripts\") pod \"keystone-db-create-s5qkh\" (UID: \"316f3a85-5cf5-44c5-b8f5-fbc6d7cbff3d\") " pod="manila-kuttl-tests/keystone-db-create-s5qkh" Feb 02 15:31:56 crc kubenswrapper[4733]: I0202 15:31:56.601866 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8g7l4\" (UniqueName: \"kubernetes.io/projected/316f3a85-5cf5-44c5-b8f5-fbc6d7cbff3d-kube-api-access-8g7l4\") pod \"keystone-db-create-s5qkh\" (UID: \"316f3a85-5cf5-44c5-b8f5-fbc6d7cbff3d\") " pod="manila-kuttl-tests/keystone-db-create-s5qkh" Feb 02 15:31:56 crc kubenswrapper[4733]: I0202 15:31:56.601925 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4jg5\" (UniqueName: \"kubernetes.io/projected/353d8d01-e2c8-40ec-80e4-034ab7786569-kube-api-access-f4jg5\") pod \"keystone-c686-account-create-update-msktt\" (UID: \"353d8d01-e2c8-40ec-80e4-034ab7786569\") " pod="manila-kuttl-tests/keystone-c686-account-create-update-msktt" Feb 02 15:31:56 crc kubenswrapper[4733]: I0202 15:31:56.601992 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/353d8d01-e2c8-40ec-80e4-034ab7786569-operator-scripts\") pod \"keystone-c686-account-create-update-msktt\" (UID: \"353d8d01-e2c8-40ec-80e4-034ab7786569\") " pod="manila-kuttl-tests/keystone-c686-account-create-update-msktt" Feb 02 15:31:56 crc kubenswrapper[4733]: I0202 15:31:56.603155 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/316f3a85-5cf5-44c5-b8f5-fbc6d7cbff3d-operator-scripts\") pod \"keystone-db-create-s5qkh\" (UID: \"316f3a85-5cf5-44c5-b8f5-fbc6d7cbff3d\") " pod="manila-kuttl-tests/keystone-db-create-s5qkh" Feb 02 15:31:56 crc kubenswrapper[4733]: I0202 15:31:56.603370 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/353d8d01-e2c8-40ec-80e4-034ab7786569-operator-scripts\") pod \"keystone-c686-account-create-update-msktt\" (UID: \"353d8d01-e2c8-40ec-80e4-034ab7786569\") " pod="manila-kuttl-tests/keystone-c686-account-create-update-msktt" Feb 02 15:31:56 crc kubenswrapper[4733]: I0202 15:31:56.633011 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4jg5\" (UniqueName: \"kubernetes.io/projected/353d8d01-e2c8-40ec-80e4-034ab7786569-kube-api-access-f4jg5\") pod \"keystone-c686-account-create-update-msktt\" (UID: \"353d8d01-e2c8-40ec-80e4-034ab7786569\") " pod="manila-kuttl-tests/keystone-c686-account-create-update-msktt" Feb 02 15:31:56 crc kubenswrapper[4733]: I0202 15:31:56.633750 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8g7l4\" (UniqueName: \"kubernetes.io/projected/316f3a85-5cf5-44c5-b8f5-fbc6d7cbff3d-kube-api-access-8g7l4\") pod \"keystone-db-create-s5qkh\" (UID: \"316f3a85-5cf5-44c5-b8f5-fbc6d7cbff3d\") " pod="manila-kuttl-tests/keystone-db-create-s5qkh" Feb 02 15:31:56 crc kubenswrapper[4733]: I0202 15:31:56.697506 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/keystone-db-create-s5qkh" Feb 02 15:31:56 crc kubenswrapper[4733]: I0202 15:31:56.703365 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["manila-kuttl-tests/ceph"] Feb 02 15:31:56 crc kubenswrapper[4733]: I0202 15:31:56.704358 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/ceph" Feb 02 15:31:56 crc kubenswrapper[4733]: I0202 15:31:56.706772 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/keystone-c686-account-create-update-msktt" Feb 02 15:31:56 crc kubenswrapper[4733]: I0202 15:31:56.707294 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"default-dockercfg-tgh29" Feb 02 15:31:56 crc kubenswrapper[4733]: I0202 15:31:56.804445 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/empty-dir/c71eca7f-f211-40bc-9116-f3c246c46e54-run\") pod \"ceph\" (UID: \"c71eca7f-f211-40bc-9116-f3c246c46e54\") " pod="manila-kuttl-tests/ceph" Feb 02 15:31:56 crc kubenswrapper[4733]: I0202 15:31:56.804511 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data\" (UniqueName: \"kubernetes.io/empty-dir/c71eca7f-f211-40bc-9116-f3c246c46e54-data\") pod \"ceph\" (UID: \"c71eca7f-f211-40bc-9116-f3c246c46e54\") " pod="manila-kuttl-tests/ceph" Feb 02 15:31:56 crc kubenswrapper[4733]: I0202 15:31:56.804534 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pc87n\" (UniqueName: \"kubernetes.io/projected/c71eca7f-f211-40bc-9116-f3c246c46e54-kube-api-access-pc87n\") pod \"ceph\" (UID: \"c71eca7f-f211-40bc-9116-f3c246c46e54\") " pod="manila-kuttl-tests/ceph" Feb 02 15:31:56 crc kubenswrapper[4733]: I0202 15:31:56.804689 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log\" (UniqueName: \"kubernetes.io/empty-dir/c71eca7f-f211-40bc-9116-f3c246c46e54-log\") pod \"ceph\" (UID: \"c71eca7f-f211-40bc-9116-f3c246c46e54\") " pod="manila-kuttl-tests/ceph" Feb 02 15:31:56 crc kubenswrapper[4733]: I0202 15:31:56.905952 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log\" (UniqueName: \"kubernetes.io/empty-dir/c71eca7f-f211-40bc-9116-f3c246c46e54-log\") pod \"ceph\" (UID: \"c71eca7f-f211-40bc-9116-f3c246c46e54\") " pod="manila-kuttl-tests/ceph" Feb 02 15:31:56 crc kubenswrapper[4733]: I0202 15:31:56.906465 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/empty-dir/c71eca7f-f211-40bc-9116-f3c246c46e54-run\") pod \"ceph\" (UID: \"c71eca7f-f211-40bc-9116-f3c246c46e54\") " pod="manila-kuttl-tests/ceph" Feb 02 15:31:56 crc kubenswrapper[4733]: I0202 15:31:56.906496 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data\" (UniqueName: \"kubernetes.io/empty-dir/c71eca7f-f211-40bc-9116-f3c246c46e54-data\") pod \"ceph\" (UID: \"c71eca7f-f211-40bc-9116-f3c246c46e54\") " pod="manila-kuttl-tests/ceph" Feb 02 15:31:56 crc kubenswrapper[4733]: I0202 15:31:56.906518 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pc87n\" (UniqueName: \"kubernetes.io/projected/c71eca7f-f211-40bc-9116-f3c246c46e54-kube-api-access-pc87n\") pod \"ceph\" (UID: \"c71eca7f-f211-40bc-9116-f3c246c46e54\") " pod="manila-kuttl-tests/ceph" Feb 02 15:31:56 crc kubenswrapper[4733]: I0202 15:31:56.906583 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log\" (UniqueName: \"kubernetes.io/empty-dir/c71eca7f-f211-40bc-9116-f3c246c46e54-log\") pod \"ceph\" (UID: \"c71eca7f-f211-40bc-9116-f3c246c46e54\") " pod="manila-kuttl-tests/ceph" Feb 02 15:31:56 crc kubenswrapper[4733]: I0202 15:31:56.906705 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/empty-dir/c71eca7f-f211-40bc-9116-f3c246c46e54-run\") pod \"ceph\" (UID: \"c71eca7f-f211-40bc-9116-f3c246c46e54\") " pod="manila-kuttl-tests/ceph" Feb 02 15:31:56 crc kubenswrapper[4733]: I0202 15:31:56.906886 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data\" (UniqueName: \"kubernetes.io/empty-dir/c71eca7f-f211-40bc-9116-f3c246c46e54-data\") pod \"ceph\" (UID: \"c71eca7f-f211-40bc-9116-f3c246c46e54\") " pod="manila-kuttl-tests/ceph" Feb 02 15:31:56 crc kubenswrapper[4733]: I0202 15:31:56.923961 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pc87n\" (UniqueName: \"kubernetes.io/projected/c71eca7f-f211-40bc-9116-f3c246c46e54-kube-api-access-pc87n\") pod \"ceph\" (UID: \"c71eca7f-f211-40bc-9116-f3c246c46e54\") " pod="manila-kuttl-tests/ceph" Feb 02 15:31:57 crc kubenswrapper[4733]: I0202 15:31:57.089606 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/ceph" Feb 02 15:31:57 crc kubenswrapper[4733]: I0202 15:31:57.154954 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/keystone-c686-account-create-update-msktt"] Feb 02 15:31:57 crc kubenswrapper[4733]: W0202 15:31:57.157715 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod353d8d01_e2c8_40ec_80e4_034ab7786569.slice/crio-d2f91a300a1841ad7ba9d8078932ae71db81162a118b5ba7388460cc41574647 WatchSource:0}: Error finding container d2f91a300a1841ad7ba9d8078932ae71db81162a118b5ba7388460cc41574647: Status 404 returned error can't find the container with id d2f91a300a1841ad7ba9d8078932ae71db81162a118b5ba7388460cc41574647 Feb 02 15:31:57 crc kubenswrapper[4733]: I0202 15:31:57.160228 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/keystone-db-create-s5qkh"] Feb 02 15:31:57 crc kubenswrapper[4733]: W0202 15:31:57.163318 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod316f3a85_5cf5_44c5_b8f5_fbc6d7cbff3d.slice/crio-f55d2a9dcd8b214b824c50ed3f29887527c1f83905a1429386b8e2936b138565 WatchSource:0}: Error finding container f55d2a9dcd8b214b824c50ed3f29887527c1f83905a1429386b8e2936b138565: Status 404 returned error can't find the container with id f55d2a9dcd8b214b824c50ed3f29887527c1f83905a1429386b8e2936b138565 Feb 02 15:31:57 crc kubenswrapper[4733]: I0202 15:31:57.636664 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/ceph" event={"ID":"c71eca7f-f211-40bc-9116-f3c246c46e54","Type":"ContainerStarted","Data":"ac9468187bfa0e717e4d13ba5017a4db3306b2c3ae4955acf6cc91b309ae3246"} Feb 02 15:31:57 crc kubenswrapper[4733]: I0202 15:31:57.638513 4733 generic.go:334] "Generic (PLEG): container finished" podID="316f3a85-5cf5-44c5-b8f5-fbc6d7cbff3d" containerID="260c837b138cab3736b1ce776c2f8f709c073d15a9f64e6dd133c80772e67431" exitCode=0 Feb 02 15:31:57 crc kubenswrapper[4733]: I0202 15:31:57.638581 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/keystone-db-create-s5qkh" event={"ID":"316f3a85-5cf5-44c5-b8f5-fbc6d7cbff3d","Type":"ContainerDied","Data":"260c837b138cab3736b1ce776c2f8f709c073d15a9f64e6dd133c80772e67431"} Feb 02 15:31:57 crc kubenswrapper[4733]: I0202 15:31:57.638616 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/keystone-db-create-s5qkh" event={"ID":"316f3a85-5cf5-44c5-b8f5-fbc6d7cbff3d","Type":"ContainerStarted","Data":"f55d2a9dcd8b214b824c50ed3f29887527c1f83905a1429386b8e2936b138565"} Feb 02 15:31:57 crc kubenswrapper[4733]: I0202 15:31:57.640515 4733 generic.go:334] "Generic (PLEG): container finished" podID="353d8d01-e2c8-40ec-80e4-034ab7786569" containerID="05672c2c2a4105cf5a9fb1aca5de5bdb6d4df5b19893e7f79cc99cc627995e50" exitCode=0 Feb 02 15:31:57 crc kubenswrapper[4733]: I0202 15:31:57.640580 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/keystone-c686-account-create-update-msktt" event={"ID":"353d8d01-e2c8-40ec-80e4-034ab7786569","Type":"ContainerDied","Data":"05672c2c2a4105cf5a9fb1aca5de5bdb6d4df5b19893e7f79cc99cc627995e50"} Feb 02 15:31:57 crc kubenswrapper[4733]: I0202 15:31:57.640619 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/keystone-c686-account-create-update-msktt" event={"ID":"353d8d01-e2c8-40ec-80e4-034ab7786569","Type":"ContainerStarted","Data":"d2f91a300a1841ad7ba9d8078932ae71db81162a118b5ba7388460cc41574647"} Feb 02 15:31:58 crc kubenswrapper[4733]: I0202 15:31:58.982822 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/keystone-c686-account-create-update-msktt" Feb 02 15:31:58 crc kubenswrapper[4733]: I0202 15:31:58.988872 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/keystone-db-create-s5qkh" Feb 02 15:31:59 crc kubenswrapper[4733]: I0202 15:31:59.136469 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/316f3a85-5cf5-44c5-b8f5-fbc6d7cbff3d-operator-scripts\") pod \"316f3a85-5cf5-44c5-b8f5-fbc6d7cbff3d\" (UID: \"316f3a85-5cf5-44c5-b8f5-fbc6d7cbff3d\") " Feb 02 15:31:59 crc kubenswrapper[4733]: I0202 15:31:59.136619 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8g7l4\" (UniqueName: \"kubernetes.io/projected/316f3a85-5cf5-44c5-b8f5-fbc6d7cbff3d-kube-api-access-8g7l4\") pod \"316f3a85-5cf5-44c5-b8f5-fbc6d7cbff3d\" (UID: \"316f3a85-5cf5-44c5-b8f5-fbc6d7cbff3d\") " Feb 02 15:31:59 crc kubenswrapper[4733]: I0202 15:31:59.136661 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f4jg5\" (UniqueName: \"kubernetes.io/projected/353d8d01-e2c8-40ec-80e4-034ab7786569-kube-api-access-f4jg5\") pod \"353d8d01-e2c8-40ec-80e4-034ab7786569\" (UID: \"353d8d01-e2c8-40ec-80e4-034ab7786569\") " Feb 02 15:31:59 crc kubenswrapper[4733]: I0202 15:31:59.136808 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/353d8d01-e2c8-40ec-80e4-034ab7786569-operator-scripts\") pod \"353d8d01-e2c8-40ec-80e4-034ab7786569\" (UID: \"353d8d01-e2c8-40ec-80e4-034ab7786569\") " Feb 02 15:31:59 crc kubenswrapper[4733]: I0202 15:31:59.137569 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/353d8d01-e2c8-40ec-80e4-034ab7786569-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "353d8d01-e2c8-40ec-80e4-034ab7786569" (UID: "353d8d01-e2c8-40ec-80e4-034ab7786569"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:31:59 crc kubenswrapper[4733]: I0202 15:31:59.137693 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/316f3a85-5cf5-44c5-b8f5-fbc6d7cbff3d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "316f3a85-5cf5-44c5-b8f5-fbc6d7cbff3d" (UID: "316f3a85-5cf5-44c5-b8f5-fbc6d7cbff3d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:31:59 crc kubenswrapper[4733]: I0202 15:31:59.138189 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/353d8d01-e2c8-40ec-80e4-034ab7786569-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 15:31:59 crc kubenswrapper[4733]: I0202 15:31:59.138222 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/316f3a85-5cf5-44c5-b8f5-fbc6d7cbff3d-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 15:31:59 crc kubenswrapper[4733]: I0202 15:31:59.142060 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/316f3a85-5cf5-44c5-b8f5-fbc6d7cbff3d-kube-api-access-8g7l4" (OuterVolumeSpecName: "kube-api-access-8g7l4") pod "316f3a85-5cf5-44c5-b8f5-fbc6d7cbff3d" (UID: "316f3a85-5cf5-44c5-b8f5-fbc6d7cbff3d"). InnerVolumeSpecName "kube-api-access-8g7l4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:31:59 crc kubenswrapper[4733]: I0202 15:31:59.142198 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/353d8d01-e2c8-40ec-80e4-034ab7786569-kube-api-access-f4jg5" (OuterVolumeSpecName: "kube-api-access-f4jg5") pod "353d8d01-e2c8-40ec-80e4-034ab7786569" (UID: "353d8d01-e2c8-40ec-80e4-034ab7786569"). InnerVolumeSpecName "kube-api-access-f4jg5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:31:59 crc kubenswrapper[4733]: I0202 15:31:59.239912 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8g7l4\" (UniqueName: \"kubernetes.io/projected/316f3a85-5cf5-44c5-b8f5-fbc6d7cbff3d-kube-api-access-8g7l4\") on node \"crc\" DevicePath \"\"" Feb 02 15:31:59 crc kubenswrapper[4733]: I0202 15:31:59.239948 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f4jg5\" (UniqueName: \"kubernetes.io/projected/353d8d01-e2c8-40ec-80e4-034ab7786569-kube-api-access-f4jg5\") on node \"crc\" DevicePath \"\"" Feb 02 15:31:59 crc kubenswrapper[4733]: I0202 15:31:59.656631 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/keystone-c686-account-create-update-msktt" event={"ID":"353d8d01-e2c8-40ec-80e4-034ab7786569","Type":"ContainerDied","Data":"d2f91a300a1841ad7ba9d8078932ae71db81162a118b5ba7388460cc41574647"} Feb 02 15:31:59 crc kubenswrapper[4733]: I0202 15:31:59.656958 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2f91a300a1841ad7ba9d8078932ae71db81162a118b5ba7388460cc41574647" Feb 02 15:31:59 crc kubenswrapper[4733]: I0202 15:31:59.657057 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/keystone-c686-account-create-update-msktt" Feb 02 15:31:59 crc kubenswrapper[4733]: I0202 15:31:59.657895 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/keystone-db-create-s5qkh" event={"ID":"316f3a85-5cf5-44c5-b8f5-fbc6d7cbff3d","Type":"ContainerDied","Data":"f55d2a9dcd8b214b824c50ed3f29887527c1f83905a1429386b8e2936b138565"} Feb 02 15:31:59 crc kubenswrapper[4733]: I0202 15:31:59.657938 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f55d2a9dcd8b214b824c50ed3f29887527c1f83905a1429386b8e2936b138565" Feb 02 15:31:59 crc kubenswrapper[4733]: I0202 15:31:59.657959 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/keystone-db-create-s5qkh" Feb 02 15:32:04 crc kubenswrapper[4733]: I0202 15:32:04.987813 4733 patch_prober.go:28] interesting pod/machine-config-daemon-zgsv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 15:32:04 crc kubenswrapper[4733]: I0202 15:32:04.988394 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 15:32:07 crc kubenswrapper[4733]: I0202 15:32:07.651342 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="manila-kuttl-tests/rabbitmq-server-0" Feb 02 15:32:08 crc kubenswrapper[4733]: I0202 15:32:08.133148 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["manila-kuttl-tests/keystone-db-sync-7wbzb"] Feb 02 15:32:08 crc kubenswrapper[4733]: E0202 15:32:08.133393 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="353d8d01-e2c8-40ec-80e4-034ab7786569" containerName="mariadb-account-create-update" Feb 02 15:32:08 crc kubenswrapper[4733]: I0202 15:32:08.133405 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="353d8d01-e2c8-40ec-80e4-034ab7786569" containerName="mariadb-account-create-update" Feb 02 15:32:08 crc kubenswrapper[4733]: E0202 15:32:08.133416 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="316f3a85-5cf5-44c5-b8f5-fbc6d7cbff3d" containerName="mariadb-database-create" Feb 02 15:32:08 crc kubenswrapper[4733]: I0202 15:32:08.133424 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="316f3a85-5cf5-44c5-b8f5-fbc6d7cbff3d" containerName="mariadb-database-create" Feb 02 15:32:08 crc kubenswrapper[4733]: I0202 15:32:08.133554 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="316f3a85-5cf5-44c5-b8f5-fbc6d7cbff3d" containerName="mariadb-database-create" Feb 02 15:32:08 crc kubenswrapper[4733]: I0202 15:32:08.133569 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="353d8d01-e2c8-40ec-80e4-034ab7786569" containerName="mariadb-account-create-update" Feb 02 15:32:08 crc kubenswrapper[4733]: I0202 15:32:08.133983 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/keystone-db-sync-7wbzb" Feb 02 15:32:08 crc kubenswrapper[4733]: I0202 15:32:08.138717 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"keystone" Feb 02 15:32:08 crc kubenswrapper[4733]: I0202 15:32:08.138985 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"keystone-config-data" Feb 02 15:32:08 crc kubenswrapper[4733]: I0202 15:32:08.139139 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"keystone-keystone-dockercfg-sq959" Feb 02 15:32:08 crc kubenswrapper[4733]: I0202 15:32:08.139179 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"keystone-scripts" Feb 02 15:32:08 crc kubenswrapper[4733]: I0202 15:32:08.192249 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/keystone-db-sync-7wbzb"] Feb 02 15:32:08 crc kubenswrapper[4733]: I0202 15:32:08.284110 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/158ef91e-84f7-4e1e-a476-569c8a755226-config-data\") pod \"keystone-db-sync-7wbzb\" (UID: \"158ef91e-84f7-4e1e-a476-569c8a755226\") " pod="manila-kuttl-tests/keystone-db-sync-7wbzb" Feb 02 15:32:08 crc kubenswrapper[4733]: I0202 15:32:08.284260 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26lwq\" (UniqueName: \"kubernetes.io/projected/158ef91e-84f7-4e1e-a476-569c8a755226-kube-api-access-26lwq\") pod \"keystone-db-sync-7wbzb\" (UID: \"158ef91e-84f7-4e1e-a476-569c8a755226\") " pod="manila-kuttl-tests/keystone-db-sync-7wbzb" Feb 02 15:32:08 crc kubenswrapper[4733]: I0202 15:32:08.385819 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26lwq\" (UniqueName: \"kubernetes.io/projected/158ef91e-84f7-4e1e-a476-569c8a755226-kube-api-access-26lwq\") pod \"keystone-db-sync-7wbzb\" (UID: \"158ef91e-84f7-4e1e-a476-569c8a755226\") " pod="manila-kuttl-tests/keystone-db-sync-7wbzb" Feb 02 15:32:08 crc kubenswrapper[4733]: I0202 15:32:08.385888 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/158ef91e-84f7-4e1e-a476-569c8a755226-config-data\") pod \"keystone-db-sync-7wbzb\" (UID: \"158ef91e-84f7-4e1e-a476-569c8a755226\") " pod="manila-kuttl-tests/keystone-db-sync-7wbzb" Feb 02 15:32:08 crc kubenswrapper[4733]: I0202 15:32:08.400312 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/158ef91e-84f7-4e1e-a476-569c8a755226-config-data\") pod \"keystone-db-sync-7wbzb\" (UID: \"158ef91e-84f7-4e1e-a476-569c8a755226\") " pod="manila-kuttl-tests/keystone-db-sync-7wbzb" Feb 02 15:32:08 crc kubenswrapper[4733]: I0202 15:32:08.403227 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26lwq\" (UniqueName: \"kubernetes.io/projected/158ef91e-84f7-4e1e-a476-569c8a755226-kube-api-access-26lwq\") pod \"keystone-db-sync-7wbzb\" (UID: \"158ef91e-84f7-4e1e-a476-569c8a755226\") " pod="manila-kuttl-tests/keystone-db-sync-7wbzb" Feb 02 15:32:08 crc kubenswrapper[4733]: I0202 15:32:08.504918 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/keystone-db-sync-7wbzb" Feb 02 15:32:09 crc kubenswrapper[4733]: I0202 15:32:09.946061 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-tqgx9"] Feb 02 15:32:09 crc kubenswrapper[4733]: I0202 15:32:09.948776 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tqgx9" Feb 02 15:32:09 crc kubenswrapper[4733]: I0202 15:32:09.971771 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tqgx9"] Feb 02 15:32:10 crc kubenswrapper[4733]: I0202 15:32:10.113134 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70d390ae-af5e-4506-bf11-bb128f8211f5-catalog-content\") pod \"certified-operators-tqgx9\" (UID: \"70d390ae-af5e-4506-bf11-bb128f8211f5\") " pod="openshift-marketplace/certified-operators-tqgx9" Feb 02 15:32:10 crc kubenswrapper[4733]: I0202 15:32:10.113218 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kk8kq\" (UniqueName: \"kubernetes.io/projected/70d390ae-af5e-4506-bf11-bb128f8211f5-kube-api-access-kk8kq\") pod \"certified-operators-tqgx9\" (UID: \"70d390ae-af5e-4506-bf11-bb128f8211f5\") " pod="openshift-marketplace/certified-operators-tqgx9" Feb 02 15:32:10 crc kubenswrapper[4733]: I0202 15:32:10.113253 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70d390ae-af5e-4506-bf11-bb128f8211f5-utilities\") pod \"certified-operators-tqgx9\" (UID: \"70d390ae-af5e-4506-bf11-bb128f8211f5\") " pod="openshift-marketplace/certified-operators-tqgx9" Feb 02 15:32:10 crc kubenswrapper[4733]: I0202 15:32:10.214660 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70d390ae-af5e-4506-bf11-bb128f8211f5-utilities\") pod \"certified-operators-tqgx9\" (UID: \"70d390ae-af5e-4506-bf11-bb128f8211f5\") " pod="openshift-marketplace/certified-operators-tqgx9" Feb 02 15:32:10 crc kubenswrapper[4733]: I0202 15:32:10.214773 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70d390ae-af5e-4506-bf11-bb128f8211f5-catalog-content\") pod \"certified-operators-tqgx9\" (UID: \"70d390ae-af5e-4506-bf11-bb128f8211f5\") " pod="openshift-marketplace/certified-operators-tqgx9" Feb 02 15:32:10 crc kubenswrapper[4733]: I0202 15:32:10.214838 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kk8kq\" (UniqueName: \"kubernetes.io/projected/70d390ae-af5e-4506-bf11-bb128f8211f5-kube-api-access-kk8kq\") pod \"certified-operators-tqgx9\" (UID: \"70d390ae-af5e-4506-bf11-bb128f8211f5\") " pod="openshift-marketplace/certified-operators-tqgx9" Feb 02 15:32:10 crc kubenswrapper[4733]: I0202 15:32:10.215411 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70d390ae-af5e-4506-bf11-bb128f8211f5-utilities\") pod \"certified-operators-tqgx9\" (UID: \"70d390ae-af5e-4506-bf11-bb128f8211f5\") " pod="openshift-marketplace/certified-operators-tqgx9" Feb 02 15:32:10 crc kubenswrapper[4733]: I0202 15:32:10.215414 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70d390ae-af5e-4506-bf11-bb128f8211f5-catalog-content\") pod \"certified-operators-tqgx9\" (UID: \"70d390ae-af5e-4506-bf11-bb128f8211f5\") " pod="openshift-marketplace/certified-operators-tqgx9" Feb 02 15:32:10 crc kubenswrapper[4733]: I0202 15:32:10.237005 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kk8kq\" (UniqueName: \"kubernetes.io/projected/70d390ae-af5e-4506-bf11-bb128f8211f5-kube-api-access-kk8kq\") pod \"certified-operators-tqgx9\" (UID: \"70d390ae-af5e-4506-bf11-bb128f8211f5\") " pod="openshift-marketplace/certified-operators-tqgx9" Feb 02 15:32:10 crc kubenswrapper[4733]: I0202 15:32:10.284761 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tqgx9" Feb 02 15:32:21 crc kubenswrapper[4733]: I0202 15:32:21.440495 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/keystone-db-sync-7wbzb"] Feb 02 15:32:21 crc kubenswrapper[4733]: I0202 15:32:21.449183 4733 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 02 15:32:21 crc kubenswrapper[4733]: E0202 15:32:21.508942 4733 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/ceph/demo:latest-squid" Feb 02 15:32:21 crc kubenswrapper[4733]: E0202 15:32:21.509384 4733 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceph,Image:quay.io/ceph/demo:latest-squid,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:MON_IP,Value:192.168.126.11,ValueFrom:nil,},EnvVar{Name:CEPH_DAEMON,Value:demo,ValueFrom:nil,},EnvVar{Name:CEPH_PUBLIC_NETWORK,Value:0.0.0.0/0,ValueFrom:nil,},EnvVar{Name:DEMO_DAEMONS,Value:osd,mds,rgw,ValueFrom:nil,},EnvVar{Name:CEPH_DEMO_UID,Value:0,ValueFrom:nil,},EnvVar{Name:RGW_NAME,Value:ceph,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:data,ReadOnly:false,MountPath:/var/lib/ceph,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:log,ReadOnly:false,MountPath:/var/log/ceph,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:run,ReadOnly:false,MountPath:/run/ceph,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pc87n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceph_manila-kuttl-tests(c71eca7f-f211-40bc-9116-f3c246c46e54): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 02 15:32:21 crc kubenswrapper[4733]: E0202 15:32:21.515533 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceph\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="manila-kuttl-tests/ceph" podUID="c71eca7f-f211-40bc-9116-f3c246c46e54" Feb 02 15:32:21 crc kubenswrapper[4733]: I0202 15:32:21.827462 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/keystone-db-sync-7wbzb" event={"ID":"158ef91e-84f7-4e1e-a476-569c8a755226","Type":"ContainerStarted","Data":"5463339a60f5e2382b15ced21642a5f6cce101de0eaa2c494bf3c60e8df7e92a"} Feb 02 15:32:21 crc kubenswrapper[4733]: E0202 15:32:21.830129 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceph\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/ceph/demo:latest-squid\\\"\"" pod="manila-kuttl-tests/ceph" podUID="c71eca7f-f211-40bc-9116-f3c246c46e54" Feb 02 15:32:21 crc kubenswrapper[4733]: I0202 15:32:21.854332 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tqgx9"] Feb 02 15:32:21 crc kubenswrapper[4733]: W0202 15:32:21.861797 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod70d390ae_af5e_4506_bf11_bb128f8211f5.slice/crio-3495ce531dc0584802bed24397ecf106e26618c7eaecb1e78bf13a84353a82a2 WatchSource:0}: Error finding container 3495ce531dc0584802bed24397ecf106e26618c7eaecb1e78bf13a84353a82a2: Status 404 returned error can't find the container with id 3495ce531dc0584802bed24397ecf106e26618c7eaecb1e78bf13a84353a82a2 Feb 02 15:32:22 crc kubenswrapper[4733]: I0202 15:32:22.834863 4733 generic.go:334] "Generic (PLEG): container finished" podID="70d390ae-af5e-4506-bf11-bb128f8211f5" containerID="44a56c4b92a7927c4cc58dc5d69b92994812aaf8871019613992f1e2ce9b865c" exitCode=0 Feb 02 15:32:22 crc kubenswrapper[4733]: I0202 15:32:22.834991 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tqgx9" event={"ID":"70d390ae-af5e-4506-bf11-bb128f8211f5","Type":"ContainerDied","Data":"44a56c4b92a7927c4cc58dc5d69b92994812aaf8871019613992f1e2ce9b865c"} Feb 02 15:32:22 crc kubenswrapper[4733]: I0202 15:32:22.835269 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tqgx9" event={"ID":"70d390ae-af5e-4506-bf11-bb128f8211f5","Type":"ContainerStarted","Data":"3495ce531dc0584802bed24397ecf106e26618c7eaecb1e78bf13a84353a82a2"} Feb 02 15:32:23 crc kubenswrapper[4733]: I0202 15:32:23.858857 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tqgx9" event={"ID":"70d390ae-af5e-4506-bf11-bb128f8211f5","Type":"ContainerStarted","Data":"44acb9e4e1f6cd23ea4602e3ac52716c4842011cb09b618057f17bca44aa3f29"} Feb 02 15:32:24 crc kubenswrapper[4733]: I0202 15:32:24.867199 4733 generic.go:334] "Generic (PLEG): container finished" podID="70d390ae-af5e-4506-bf11-bb128f8211f5" containerID="44acb9e4e1f6cd23ea4602e3ac52716c4842011cb09b618057f17bca44aa3f29" exitCode=0 Feb 02 15:32:24 crc kubenswrapper[4733]: I0202 15:32:24.867255 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tqgx9" event={"ID":"70d390ae-af5e-4506-bf11-bb128f8211f5","Type":"ContainerDied","Data":"44acb9e4e1f6cd23ea4602e3ac52716c4842011cb09b618057f17bca44aa3f29"} Feb 02 15:32:32 crc kubenswrapper[4733]: I0202 15:32:32.929213 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tqgx9" event={"ID":"70d390ae-af5e-4506-bf11-bb128f8211f5","Type":"ContainerStarted","Data":"2627f4dceff09d170e28b839eb51144849429c7257f430b16aec60e4804a0fbb"} Feb 02 15:32:32 crc kubenswrapper[4733]: I0202 15:32:32.931455 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/keystone-db-sync-7wbzb" event={"ID":"158ef91e-84f7-4e1e-a476-569c8a755226","Type":"ContainerStarted","Data":"0eb75ac47796e3d56aedc56514b0b53601c1dcf4521b4e81651ca5e589242b0e"} Feb 02 15:32:32 crc kubenswrapper[4733]: I0202 15:32:32.965779 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-tqgx9" podStartSLOduration=14.7052812 podStartE2EDuration="23.965760909s" podCreationTimestamp="2026-02-02 15:32:09 +0000 UTC" firstStartedPulling="2026-02-02 15:32:22.836256769 +0000 UTC m=+1026.287718127" lastFinishedPulling="2026-02-02 15:32:32.096736438 +0000 UTC m=+1035.548197836" observedRunningTime="2026-02-02 15:32:32.960848689 +0000 UTC m=+1036.412310047" watchObservedRunningTime="2026-02-02 15:32:32.965760909 +0000 UTC m=+1036.417222267" Feb 02 15:32:32 crc kubenswrapper[4733]: I0202 15:32:32.981021 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="manila-kuttl-tests/keystone-db-sync-7wbzb" podStartSLOduration=14.304713983 podStartE2EDuration="24.980999614s" podCreationTimestamp="2026-02-02 15:32:08 +0000 UTC" firstStartedPulling="2026-02-02 15:32:21.448873241 +0000 UTC m=+1024.900334599" lastFinishedPulling="2026-02-02 15:32:32.125158842 +0000 UTC m=+1035.576620230" observedRunningTime="2026-02-02 15:32:32.978935119 +0000 UTC m=+1036.430396487" watchObservedRunningTime="2026-02-02 15:32:32.980999614 +0000 UTC m=+1036.432460992" Feb 02 15:32:34 crc kubenswrapper[4733]: I0202 15:32:34.988263 4733 patch_prober.go:28] interesting pod/machine-config-daemon-zgsv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 15:32:34 crc kubenswrapper[4733]: I0202 15:32:34.989381 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 15:32:40 crc kubenswrapper[4733]: I0202 15:32:40.285215 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-tqgx9" Feb 02 15:32:40 crc kubenswrapper[4733]: I0202 15:32:40.285622 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-tqgx9" Feb 02 15:32:40 crc kubenswrapper[4733]: I0202 15:32:40.368835 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-tqgx9" Feb 02 15:32:41 crc kubenswrapper[4733]: I0202 15:32:41.076965 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-tqgx9" Feb 02 15:32:41 crc kubenswrapper[4733]: I0202 15:32:41.172535 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tqgx9"] Feb 02 15:32:43 crc kubenswrapper[4733]: I0202 15:32:43.017241 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-tqgx9" podUID="70d390ae-af5e-4506-bf11-bb128f8211f5" containerName="registry-server" containerID="cri-o://2627f4dceff09d170e28b839eb51144849429c7257f430b16aec60e4804a0fbb" gracePeriod=2 Feb 02 15:32:44 crc kubenswrapper[4733]: I0202 15:32:44.030101 4733 generic.go:334] "Generic (PLEG): container finished" podID="70d390ae-af5e-4506-bf11-bb128f8211f5" containerID="2627f4dceff09d170e28b839eb51144849429c7257f430b16aec60e4804a0fbb" exitCode=0 Feb 02 15:32:44 crc kubenswrapper[4733]: I0202 15:32:44.030235 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tqgx9" event={"ID":"70d390ae-af5e-4506-bf11-bb128f8211f5","Type":"ContainerDied","Data":"2627f4dceff09d170e28b839eb51144849429c7257f430b16aec60e4804a0fbb"} Feb 02 15:32:45 crc kubenswrapper[4733]: I0202 15:32:45.587376 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tqgx9" Feb 02 15:32:45 crc kubenswrapper[4733]: I0202 15:32:45.693025 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kk8kq\" (UniqueName: \"kubernetes.io/projected/70d390ae-af5e-4506-bf11-bb128f8211f5-kube-api-access-kk8kq\") pod \"70d390ae-af5e-4506-bf11-bb128f8211f5\" (UID: \"70d390ae-af5e-4506-bf11-bb128f8211f5\") " Feb 02 15:32:45 crc kubenswrapper[4733]: I0202 15:32:45.693102 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70d390ae-af5e-4506-bf11-bb128f8211f5-utilities\") pod \"70d390ae-af5e-4506-bf11-bb128f8211f5\" (UID: \"70d390ae-af5e-4506-bf11-bb128f8211f5\") " Feb 02 15:32:45 crc kubenswrapper[4733]: I0202 15:32:45.693137 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70d390ae-af5e-4506-bf11-bb128f8211f5-catalog-content\") pod \"70d390ae-af5e-4506-bf11-bb128f8211f5\" (UID: \"70d390ae-af5e-4506-bf11-bb128f8211f5\") " Feb 02 15:32:45 crc kubenswrapper[4733]: I0202 15:32:45.693909 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70d390ae-af5e-4506-bf11-bb128f8211f5-utilities" (OuterVolumeSpecName: "utilities") pod "70d390ae-af5e-4506-bf11-bb128f8211f5" (UID: "70d390ae-af5e-4506-bf11-bb128f8211f5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:32:45 crc kubenswrapper[4733]: I0202 15:32:45.705323 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70d390ae-af5e-4506-bf11-bb128f8211f5-kube-api-access-kk8kq" (OuterVolumeSpecName: "kube-api-access-kk8kq") pod "70d390ae-af5e-4506-bf11-bb128f8211f5" (UID: "70d390ae-af5e-4506-bf11-bb128f8211f5"). InnerVolumeSpecName "kube-api-access-kk8kq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:32:45 crc kubenswrapper[4733]: I0202 15:32:45.771871 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70d390ae-af5e-4506-bf11-bb128f8211f5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "70d390ae-af5e-4506-bf11-bb128f8211f5" (UID: "70d390ae-af5e-4506-bf11-bb128f8211f5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:32:45 crc kubenswrapper[4733]: I0202 15:32:45.794846 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70d390ae-af5e-4506-bf11-bb128f8211f5-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 15:32:45 crc kubenswrapper[4733]: I0202 15:32:45.794887 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kk8kq\" (UniqueName: \"kubernetes.io/projected/70d390ae-af5e-4506-bf11-bb128f8211f5-kube-api-access-kk8kq\") on node \"crc\" DevicePath \"\"" Feb 02 15:32:45 crc kubenswrapper[4733]: I0202 15:32:45.794898 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70d390ae-af5e-4506-bf11-bb128f8211f5-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 15:32:46 crc kubenswrapper[4733]: I0202 15:32:46.046107 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tqgx9" Feb 02 15:32:46 crc kubenswrapper[4733]: I0202 15:32:46.046099 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tqgx9" event={"ID":"70d390ae-af5e-4506-bf11-bb128f8211f5","Type":"ContainerDied","Data":"3495ce531dc0584802bed24397ecf106e26618c7eaecb1e78bf13a84353a82a2"} Feb 02 15:32:46 crc kubenswrapper[4733]: I0202 15:32:46.046215 4733 scope.go:117] "RemoveContainer" containerID="2627f4dceff09d170e28b839eb51144849429c7257f430b16aec60e4804a0fbb" Feb 02 15:32:46 crc kubenswrapper[4733]: I0202 15:32:46.047730 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/ceph" event={"ID":"c71eca7f-f211-40bc-9116-f3c246c46e54","Type":"ContainerStarted","Data":"00f13c6c3710736a5fd10709601532b0ab4f0709341f931f4399424374017bce"} Feb 02 15:32:46 crc kubenswrapper[4733]: I0202 15:32:46.068439 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="manila-kuttl-tests/ceph" podStartSLOduration=5.80080568 podStartE2EDuration="50.068419172s" podCreationTimestamp="2026-02-02 15:31:56 +0000 UTC" firstStartedPulling="2026-02-02 15:31:57.116011219 +0000 UTC m=+1000.567472577" lastFinishedPulling="2026-02-02 15:32:41.383624671 +0000 UTC m=+1044.835086069" observedRunningTime="2026-02-02 15:32:46.065325241 +0000 UTC m=+1049.516786639" watchObservedRunningTime="2026-02-02 15:32:46.068419172 +0000 UTC m=+1049.519880530" Feb 02 15:32:46 crc kubenswrapper[4733]: I0202 15:32:46.076062 4733 scope.go:117] "RemoveContainer" containerID="44acb9e4e1f6cd23ea4602e3ac52716c4842011cb09b618057f17bca44aa3f29" Feb 02 15:32:46 crc kubenswrapper[4733]: I0202 15:32:46.086634 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tqgx9"] Feb 02 15:32:46 crc kubenswrapper[4733]: I0202 15:32:46.092517 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-tqgx9"] Feb 02 15:32:46 crc kubenswrapper[4733]: I0202 15:32:46.101640 4733 scope.go:117] "RemoveContainer" containerID="44a56c4b92a7927c4cc58dc5d69b92994812aaf8871019613992f1e2ce9b865c" Feb 02 15:32:47 crc kubenswrapper[4733]: I0202 15:32:47.057411 4733 generic.go:334] "Generic (PLEG): container finished" podID="158ef91e-84f7-4e1e-a476-569c8a755226" containerID="0eb75ac47796e3d56aedc56514b0b53601c1dcf4521b4e81651ca5e589242b0e" exitCode=0 Feb 02 15:32:47 crc kubenswrapper[4733]: I0202 15:32:47.057489 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/keystone-db-sync-7wbzb" event={"ID":"158ef91e-84f7-4e1e-a476-569c8a755226","Type":"ContainerDied","Data":"0eb75ac47796e3d56aedc56514b0b53601c1dcf4521b4e81651ca5e589242b0e"} Feb 02 15:32:47 crc kubenswrapper[4733]: I0202 15:32:47.263764 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70d390ae-af5e-4506-bf11-bb128f8211f5" path="/var/lib/kubelet/pods/70d390ae-af5e-4506-bf11-bb128f8211f5/volumes" Feb 02 15:32:48 crc kubenswrapper[4733]: I0202 15:32:48.383347 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/keystone-db-sync-7wbzb" Feb 02 15:32:48 crc kubenswrapper[4733]: I0202 15:32:48.534094 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/158ef91e-84f7-4e1e-a476-569c8a755226-config-data\") pod \"158ef91e-84f7-4e1e-a476-569c8a755226\" (UID: \"158ef91e-84f7-4e1e-a476-569c8a755226\") " Feb 02 15:32:48 crc kubenswrapper[4733]: I0202 15:32:48.534343 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-26lwq\" (UniqueName: \"kubernetes.io/projected/158ef91e-84f7-4e1e-a476-569c8a755226-kube-api-access-26lwq\") pod \"158ef91e-84f7-4e1e-a476-569c8a755226\" (UID: \"158ef91e-84f7-4e1e-a476-569c8a755226\") " Feb 02 15:32:48 crc kubenswrapper[4733]: I0202 15:32:48.539951 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/158ef91e-84f7-4e1e-a476-569c8a755226-kube-api-access-26lwq" (OuterVolumeSpecName: "kube-api-access-26lwq") pod "158ef91e-84f7-4e1e-a476-569c8a755226" (UID: "158ef91e-84f7-4e1e-a476-569c8a755226"). InnerVolumeSpecName "kube-api-access-26lwq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:32:48 crc kubenswrapper[4733]: I0202 15:32:48.597842 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/158ef91e-84f7-4e1e-a476-569c8a755226-config-data" (OuterVolumeSpecName: "config-data") pod "158ef91e-84f7-4e1e-a476-569c8a755226" (UID: "158ef91e-84f7-4e1e-a476-569c8a755226"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:32:48 crc kubenswrapper[4733]: I0202 15:32:48.637097 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/158ef91e-84f7-4e1e-a476-569c8a755226-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 15:32:48 crc kubenswrapper[4733]: I0202 15:32:48.637149 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-26lwq\" (UniqueName: \"kubernetes.io/projected/158ef91e-84f7-4e1e-a476-569c8a755226-kube-api-access-26lwq\") on node \"crc\" DevicePath \"\"" Feb 02 15:32:49 crc kubenswrapper[4733]: I0202 15:32:49.075635 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/keystone-db-sync-7wbzb" event={"ID":"158ef91e-84f7-4e1e-a476-569c8a755226","Type":"ContainerDied","Data":"5463339a60f5e2382b15ced21642a5f6cce101de0eaa2c494bf3c60e8df7e92a"} Feb 02 15:32:49 crc kubenswrapper[4733]: I0202 15:32:49.075969 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5463339a60f5e2382b15ced21642a5f6cce101de0eaa2c494bf3c60e8df7e92a" Feb 02 15:32:49 crc kubenswrapper[4733]: I0202 15:32:49.075750 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/keystone-db-sync-7wbzb" Feb 02 15:32:49 crc kubenswrapper[4733]: I0202 15:32:49.306900 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["manila-kuttl-tests/keystone-bootstrap-65qw9"] Feb 02 15:32:49 crc kubenswrapper[4733]: E0202 15:32:49.307130 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70d390ae-af5e-4506-bf11-bb128f8211f5" containerName="extract-content" Feb 02 15:32:49 crc kubenswrapper[4733]: I0202 15:32:49.307142 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="70d390ae-af5e-4506-bf11-bb128f8211f5" containerName="extract-content" Feb 02 15:32:49 crc kubenswrapper[4733]: E0202 15:32:49.307149 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="158ef91e-84f7-4e1e-a476-569c8a755226" containerName="keystone-db-sync" Feb 02 15:32:49 crc kubenswrapper[4733]: I0202 15:32:49.307154 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="158ef91e-84f7-4e1e-a476-569c8a755226" containerName="keystone-db-sync" Feb 02 15:32:49 crc kubenswrapper[4733]: E0202 15:32:49.307186 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70d390ae-af5e-4506-bf11-bb128f8211f5" containerName="extract-utilities" Feb 02 15:32:49 crc kubenswrapper[4733]: I0202 15:32:49.307193 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="70d390ae-af5e-4506-bf11-bb128f8211f5" containerName="extract-utilities" Feb 02 15:32:49 crc kubenswrapper[4733]: E0202 15:32:49.307204 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70d390ae-af5e-4506-bf11-bb128f8211f5" containerName="registry-server" Feb 02 15:32:49 crc kubenswrapper[4733]: I0202 15:32:49.307210 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="70d390ae-af5e-4506-bf11-bb128f8211f5" containerName="registry-server" Feb 02 15:32:49 crc kubenswrapper[4733]: I0202 15:32:49.307316 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="158ef91e-84f7-4e1e-a476-569c8a755226" containerName="keystone-db-sync" Feb 02 15:32:49 crc kubenswrapper[4733]: I0202 15:32:49.307334 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="70d390ae-af5e-4506-bf11-bb128f8211f5" containerName="registry-server" Feb 02 15:32:49 crc kubenswrapper[4733]: I0202 15:32:49.307709 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/keystone-bootstrap-65qw9" Feb 02 15:32:49 crc kubenswrapper[4733]: I0202 15:32:49.310387 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"osp-secret" Feb 02 15:32:49 crc kubenswrapper[4733]: I0202 15:32:49.310709 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"keystone-keystone-dockercfg-sq959" Feb 02 15:32:49 crc kubenswrapper[4733]: I0202 15:32:49.310745 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"keystone-config-data" Feb 02 15:32:49 crc kubenswrapper[4733]: I0202 15:32:49.311357 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"keystone-scripts" Feb 02 15:32:49 crc kubenswrapper[4733]: I0202 15:32:49.313237 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"keystone" Feb 02 15:32:49 crc kubenswrapper[4733]: I0202 15:32:49.385501 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/keystone-bootstrap-65qw9"] Feb 02 15:32:49 crc kubenswrapper[4733]: I0202 15:32:49.448770 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21e8b34c-9313-4546-8c0d-0b3b014e190c-scripts\") pod \"keystone-bootstrap-65qw9\" (UID: \"21e8b34c-9313-4546-8c0d-0b3b014e190c\") " pod="manila-kuttl-tests/keystone-bootstrap-65qw9" Feb 02 15:32:49 crc kubenswrapper[4733]: I0202 15:32:49.449325 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/21e8b34c-9313-4546-8c0d-0b3b014e190c-credential-keys\") pod \"keystone-bootstrap-65qw9\" (UID: \"21e8b34c-9313-4546-8c0d-0b3b014e190c\") " pod="manila-kuttl-tests/keystone-bootstrap-65qw9" Feb 02 15:32:49 crc kubenswrapper[4733]: I0202 15:32:49.449405 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/21e8b34c-9313-4546-8c0d-0b3b014e190c-fernet-keys\") pod \"keystone-bootstrap-65qw9\" (UID: \"21e8b34c-9313-4546-8c0d-0b3b014e190c\") " pod="manila-kuttl-tests/keystone-bootstrap-65qw9" Feb 02 15:32:49 crc kubenswrapper[4733]: I0202 15:32:49.449576 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21e8b34c-9313-4546-8c0d-0b3b014e190c-config-data\") pod \"keystone-bootstrap-65qw9\" (UID: \"21e8b34c-9313-4546-8c0d-0b3b014e190c\") " pod="manila-kuttl-tests/keystone-bootstrap-65qw9" Feb 02 15:32:49 crc kubenswrapper[4733]: I0202 15:32:49.449617 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2b6n\" (UniqueName: \"kubernetes.io/projected/21e8b34c-9313-4546-8c0d-0b3b014e190c-kube-api-access-p2b6n\") pod \"keystone-bootstrap-65qw9\" (UID: \"21e8b34c-9313-4546-8c0d-0b3b014e190c\") " pod="manila-kuttl-tests/keystone-bootstrap-65qw9" Feb 02 15:32:49 crc kubenswrapper[4733]: I0202 15:32:49.551359 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21e8b34c-9313-4546-8c0d-0b3b014e190c-config-data\") pod \"keystone-bootstrap-65qw9\" (UID: \"21e8b34c-9313-4546-8c0d-0b3b014e190c\") " pod="manila-kuttl-tests/keystone-bootstrap-65qw9" Feb 02 15:32:49 crc kubenswrapper[4733]: I0202 15:32:49.552907 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2b6n\" (UniqueName: \"kubernetes.io/projected/21e8b34c-9313-4546-8c0d-0b3b014e190c-kube-api-access-p2b6n\") pod \"keystone-bootstrap-65qw9\" (UID: \"21e8b34c-9313-4546-8c0d-0b3b014e190c\") " pod="manila-kuttl-tests/keystone-bootstrap-65qw9" Feb 02 15:32:49 crc kubenswrapper[4733]: I0202 15:32:49.552966 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21e8b34c-9313-4546-8c0d-0b3b014e190c-scripts\") pod \"keystone-bootstrap-65qw9\" (UID: \"21e8b34c-9313-4546-8c0d-0b3b014e190c\") " pod="manila-kuttl-tests/keystone-bootstrap-65qw9" Feb 02 15:32:49 crc kubenswrapper[4733]: I0202 15:32:49.553119 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/21e8b34c-9313-4546-8c0d-0b3b014e190c-credential-keys\") pod \"keystone-bootstrap-65qw9\" (UID: \"21e8b34c-9313-4546-8c0d-0b3b014e190c\") " pod="manila-kuttl-tests/keystone-bootstrap-65qw9" Feb 02 15:32:49 crc kubenswrapper[4733]: I0202 15:32:49.553199 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/21e8b34c-9313-4546-8c0d-0b3b014e190c-fernet-keys\") pod \"keystone-bootstrap-65qw9\" (UID: \"21e8b34c-9313-4546-8c0d-0b3b014e190c\") " pod="manila-kuttl-tests/keystone-bootstrap-65qw9" Feb 02 15:32:49 crc kubenswrapper[4733]: I0202 15:32:49.556268 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/21e8b34c-9313-4546-8c0d-0b3b014e190c-credential-keys\") pod \"keystone-bootstrap-65qw9\" (UID: \"21e8b34c-9313-4546-8c0d-0b3b014e190c\") " pod="manila-kuttl-tests/keystone-bootstrap-65qw9" Feb 02 15:32:49 crc kubenswrapper[4733]: I0202 15:32:49.556276 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21e8b34c-9313-4546-8c0d-0b3b014e190c-scripts\") pod \"keystone-bootstrap-65qw9\" (UID: \"21e8b34c-9313-4546-8c0d-0b3b014e190c\") " pod="manila-kuttl-tests/keystone-bootstrap-65qw9" Feb 02 15:32:49 crc kubenswrapper[4733]: I0202 15:32:49.557105 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21e8b34c-9313-4546-8c0d-0b3b014e190c-config-data\") pod \"keystone-bootstrap-65qw9\" (UID: \"21e8b34c-9313-4546-8c0d-0b3b014e190c\") " pod="manila-kuttl-tests/keystone-bootstrap-65qw9" Feb 02 15:32:49 crc kubenswrapper[4733]: I0202 15:32:49.557970 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/21e8b34c-9313-4546-8c0d-0b3b014e190c-fernet-keys\") pod \"keystone-bootstrap-65qw9\" (UID: \"21e8b34c-9313-4546-8c0d-0b3b014e190c\") " pod="manila-kuttl-tests/keystone-bootstrap-65qw9" Feb 02 15:32:49 crc kubenswrapper[4733]: I0202 15:32:49.591327 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2b6n\" (UniqueName: \"kubernetes.io/projected/21e8b34c-9313-4546-8c0d-0b3b014e190c-kube-api-access-p2b6n\") pod \"keystone-bootstrap-65qw9\" (UID: \"21e8b34c-9313-4546-8c0d-0b3b014e190c\") " pod="manila-kuttl-tests/keystone-bootstrap-65qw9" Feb 02 15:32:49 crc kubenswrapper[4733]: I0202 15:32:49.627020 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/keystone-bootstrap-65qw9" Feb 02 15:32:50 crc kubenswrapper[4733]: I0202 15:32:50.080790 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/keystone-bootstrap-65qw9"] Feb 02 15:32:50 crc kubenswrapper[4733]: W0202 15:32:50.082715 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod21e8b34c_9313_4546_8c0d_0b3b014e190c.slice/crio-cad26bd87198d44bad5435c8aedc1aa045fa79005e67c3c81bbf854b348e1675 WatchSource:0}: Error finding container cad26bd87198d44bad5435c8aedc1aa045fa79005e67c3c81bbf854b348e1675: Status 404 returned error can't find the container with id cad26bd87198d44bad5435c8aedc1aa045fa79005e67c3c81bbf854b348e1675 Feb 02 15:32:51 crc kubenswrapper[4733]: I0202 15:32:51.099847 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/keystone-bootstrap-65qw9" event={"ID":"21e8b34c-9313-4546-8c0d-0b3b014e190c","Type":"ContainerStarted","Data":"8039566f2e14aac9a5310744a17130b451c216905c95273311f102f4d823d38e"} Feb 02 15:32:51 crc kubenswrapper[4733]: I0202 15:32:51.100388 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/keystone-bootstrap-65qw9" event={"ID":"21e8b34c-9313-4546-8c0d-0b3b014e190c","Type":"ContainerStarted","Data":"cad26bd87198d44bad5435c8aedc1aa045fa79005e67c3c81bbf854b348e1675"} Feb 02 15:32:51 crc kubenswrapper[4733]: I0202 15:32:51.120718 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="manila-kuttl-tests/keystone-bootstrap-65qw9" podStartSLOduration=2.120687491 podStartE2EDuration="2.120687491s" podCreationTimestamp="2026-02-02 15:32:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:32:51.118655708 +0000 UTC m=+1054.570117116" watchObservedRunningTime="2026-02-02 15:32:51.120687491 +0000 UTC m=+1054.572148889" Feb 02 15:32:53 crc kubenswrapper[4733]: I0202 15:32:53.115998 4733 generic.go:334] "Generic (PLEG): container finished" podID="21e8b34c-9313-4546-8c0d-0b3b014e190c" containerID="8039566f2e14aac9a5310744a17130b451c216905c95273311f102f4d823d38e" exitCode=0 Feb 02 15:32:53 crc kubenswrapper[4733]: I0202 15:32:53.116133 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/keystone-bootstrap-65qw9" event={"ID":"21e8b34c-9313-4546-8c0d-0b3b014e190c","Type":"ContainerDied","Data":"8039566f2e14aac9a5310744a17130b451c216905c95273311f102f4d823d38e"} Feb 02 15:32:54 crc kubenswrapper[4733]: I0202 15:32:54.413593 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/keystone-bootstrap-65qw9" Feb 02 15:32:54 crc kubenswrapper[4733]: I0202 15:32:54.545148 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21e8b34c-9313-4546-8c0d-0b3b014e190c-scripts\") pod \"21e8b34c-9313-4546-8c0d-0b3b014e190c\" (UID: \"21e8b34c-9313-4546-8c0d-0b3b014e190c\") " Feb 02 15:32:54 crc kubenswrapper[4733]: I0202 15:32:54.545262 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/21e8b34c-9313-4546-8c0d-0b3b014e190c-credential-keys\") pod \"21e8b34c-9313-4546-8c0d-0b3b014e190c\" (UID: \"21e8b34c-9313-4546-8c0d-0b3b014e190c\") " Feb 02 15:32:54 crc kubenswrapper[4733]: I0202 15:32:54.545316 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/21e8b34c-9313-4546-8c0d-0b3b014e190c-fernet-keys\") pod \"21e8b34c-9313-4546-8c0d-0b3b014e190c\" (UID: \"21e8b34c-9313-4546-8c0d-0b3b014e190c\") " Feb 02 15:32:54 crc kubenswrapper[4733]: I0202 15:32:54.545370 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21e8b34c-9313-4546-8c0d-0b3b014e190c-config-data\") pod \"21e8b34c-9313-4546-8c0d-0b3b014e190c\" (UID: \"21e8b34c-9313-4546-8c0d-0b3b014e190c\") " Feb 02 15:32:54 crc kubenswrapper[4733]: I0202 15:32:54.545467 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p2b6n\" (UniqueName: \"kubernetes.io/projected/21e8b34c-9313-4546-8c0d-0b3b014e190c-kube-api-access-p2b6n\") pod \"21e8b34c-9313-4546-8c0d-0b3b014e190c\" (UID: \"21e8b34c-9313-4546-8c0d-0b3b014e190c\") " Feb 02 15:32:54 crc kubenswrapper[4733]: I0202 15:32:54.568614 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21e8b34c-9313-4546-8c0d-0b3b014e190c-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "21e8b34c-9313-4546-8c0d-0b3b014e190c" (UID: "21e8b34c-9313-4546-8c0d-0b3b014e190c"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:32:54 crc kubenswrapper[4733]: I0202 15:32:54.568680 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21e8b34c-9313-4546-8c0d-0b3b014e190c-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "21e8b34c-9313-4546-8c0d-0b3b014e190c" (UID: "21e8b34c-9313-4546-8c0d-0b3b014e190c"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:32:54 crc kubenswrapper[4733]: I0202 15:32:54.568715 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21e8b34c-9313-4546-8c0d-0b3b014e190c-kube-api-access-p2b6n" (OuterVolumeSpecName: "kube-api-access-p2b6n") pod "21e8b34c-9313-4546-8c0d-0b3b014e190c" (UID: "21e8b34c-9313-4546-8c0d-0b3b014e190c"). InnerVolumeSpecName "kube-api-access-p2b6n". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:32:54 crc kubenswrapper[4733]: I0202 15:32:54.568718 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21e8b34c-9313-4546-8c0d-0b3b014e190c-scripts" (OuterVolumeSpecName: "scripts") pod "21e8b34c-9313-4546-8c0d-0b3b014e190c" (UID: "21e8b34c-9313-4546-8c0d-0b3b014e190c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:32:54 crc kubenswrapper[4733]: I0202 15:32:54.584396 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21e8b34c-9313-4546-8c0d-0b3b014e190c-config-data" (OuterVolumeSpecName: "config-data") pod "21e8b34c-9313-4546-8c0d-0b3b014e190c" (UID: "21e8b34c-9313-4546-8c0d-0b3b014e190c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:32:54 crc kubenswrapper[4733]: I0202 15:32:54.647189 4733 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/21e8b34c-9313-4546-8c0d-0b3b014e190c-credential-keys\") on node \"crc\" DevicePath \"\"" Feb 02 15:32:54 crc kubenswrapper[4733]: I0202 15:32:54.647253 4733 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/21e8b34c-9313-4546-8c0d-0b3b014e190c-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 02 15:32:54 crc kubenswrapper[4733]: I0202 15:32:54.647272 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21e8b34c-9313-4546-8c0d-0b3b014e190c-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 15:32:54 crc kubenswrapper[4733]: I0202 15:32:54.647291 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p2b6n\" (UniqueName: \"kubernetes.io/projected/21e8b34c-9313-4546-8c0d-0b3b014e190c-kube-api-access-p2b6n\") on node \"crc\" DevicePath \"\"" Feb 02 15:32:54 crc kubenswrapper[4733]: I0202 15:32:54.647346 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21e8b34c-9313-4546-8c0d-0b3b014e190c-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 15:32:55 crc kubenswrapper[4733]: I0202 15:32:55.137723 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/keystone-bootstrap-65qw9" event={"ID":"21e8b34c-9313-4546-8c0d-0b3b014e190c","Type":"ContainerDied","Data":"cad26bd87198d44bad5435c8aedc1aa045fa79005e67c3c81bbf854b348e1675"} Feb 02 15:32:55 crc kubenswrapper[4733]: I0202 15:32:55.137776 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cad26bd87198d44bad5435c8aedc1aa045fa79005e67c3c81bbf854b348e1675" Feb 02 15:32:55 crc kubenswrapper[4733]: I0202 15:32:55.137803 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/keystone-bootstrap-65qw9" Feb 02 15:32:55 crc kubenswrapper[4733]: I0202 15:32:55.219375 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["manila-kuttl-tests/keystone-5f5c86bdc8-jpphg"] Feb 02 15:32:55 crc kubenswrapper[4733]: E0202 15:32:55.219788 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21e8b34c-9313-4546-8c0d-0b3b014e190c" containerName="keystone-bootstrap" Feb 02 15:32:55 crc kubenswrapper[4733]: I0202 15:32:55.219879 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="21e8b34c-9313-4546-8c0d-0b3b014e190c" containerName="keystone-bootstrap" Feb 02 15:32:55 crc kubenswrapper[4733]: I0202 15:32:55.220053 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="21e8b34c-9313-4546-8c0d-0b3b014e190c" containerName="keystone-bootstrap" Feb 02 15:32:55 crc kubenswrapper[4733]: I0202 15:32:55.220542 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/keystone-5f5c86bdc8-jpphg" Feb 02 15:32:55 crc kubenswrapper[4733]: I0202 15:32:55.222768 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"keystone-keystone-dockercfg-sq959" Feb 02 15:32:55 crc kubenswrapper[4733]: I0202 15:32:55.224397 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"keystone-config-data" Feb 02 15:32:55 crc kubenswrapper[4733]: I0202 15:32:55.224547 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"keystone-scripts" Feb 02 15:32:55 crc kubenswrapper[4733]: I0202 15:32:55.224736 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"keystone" Feb 02 15:32:55 crc kubenswrapper[4733]: I0202 15:32:55.246126 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/keystone-5f5c86bdc8-jpphg"] Feb 02 15:32:55 crc kubenswrapper[4733]: I0202 15:32:55.359405 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nr7m\" (UniqueName: \"kubernetes.io/projected/a972e8cd-cae8-4894-b2d3-7bfff75acc1c-kube-api-access-5nr7m\") pod \"keystone-5f5c86bdc8-jpphg\" (UID: \"a972e8cd-cae8-4894-b2d3-7bfff75acc1c\") " pod="manila-kuttl-tests/keystone-5f5c86bdc8-jpphg" Feb 02 15:32:55 crc kubenswrapper[4733]: I0202 15:32:55.359466 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a972e8cd-cae8-4894-b2d3-7bfff75acc1c-config-data\") pod \"keystone-5f5c86bdc8-jpphg\" (UID: \"a972e8cd-cae8-4894-b2d3-7bfff75acc1c\") " pod="manila-kuttl-tests/keystone-5f5c86bdc8-jpphg" Feb 02 15:32:55 crc kubenswrapper[4733]: I0202 15:32:55.359771 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a972e8cd-cae8-4894-b2d3-7bfff75acc1c-fernet-keys\") pod \"keystone-5f5c86bdc8-jpphg\" (UID: \"a972e8cd-cae8-4894-b2d3-7bfff75acc1c\") " pod="manila-kuttl-tests/keystone-5f5c86bdc8-jpphg" Feb 02 15:32:55 crc kubenswrapper[4733]: I0202 15:32:55.359812 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a972e8cd-cae8-4894-b2d3-7bfff75acc1c-scripts\") pod \"keystone-5f5c86bdc8-jpphg\" (UID: \"a972e8cd-cae8-4894-b2d3-7bfff75acc1c\") " pod="manila-kuttl-tests/keystone-5f5c86bdc8-jpphg" Feb 02 15:32:55 crc kubenswrapper[4733]: I0202 15:32:55.359874 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a972e8cd-cae8-4894-b2d3-7bfff75acc1c-credential-keys\") pod \"keystone-5f5c86bdc8-jpphg\" (UID: \"a972e8cd-cae8-4894-b2d3-7bfff75acc1c\") " pod="manila-kuttl-tests/keystone-5f5c86bdc8-jpphg" Feb 02 15:32:55 crc kubenswrapper[4733]: I0202 15:32:55.461817 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a972e8cd-cae8-4894-b2d3-7bfff75acc1c-fernet-keys\") pod \"keystone-5f5c86bdc8-jpphg\" (UID: \"a972e8cd-cae8-4894-b2d3-7bfff75acc1c\") " pod="manila-kuttl-tests/keystone-5f5c86bdc8-jpphg" Feb 02 15:32:55 crc kubenswrapper[4733]: I0202 15:32:55.461922 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a972e8cd-cae8-4894-b2d3-7bfff75acc1c-scripts\") pod \"keystone-5f5c86bdc8-jpphg\" (UID: \"a972e8cd-cae8-4894-b2d3-7bfff75acc1c\") " pod="manila-kuttl-tests/keystone-5f5c86bdc8-jpphg" Feb 02 15:32:55 crc kubenswrapper[4733]: I0202 15:32:55.462002 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a972e8cd-cae8-4894-b2d3-7bfff75acc1c-credential-keys\") pod \"keystone-5f5c86bdc8-jpphg\" (UID: \"a972e8cd-cae8-4894-b2d3-7bfff75acc1c\") " pod="manila-kuttl-tests/keystone-5f5c86bdc8-jpphg" Feb 02 15:32:55 crc kubenswrapper[4733]: I0202 15:32:55.462100 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nr7m\" (UniqueName: \"kubernetes.io/projected/a972e8cd-cae8-4894-b2d3-7bfff75acc1c-kube-api-access-5nr7m\") pod \"keystone-5f5c86bdc8-jpphg\" (UID: \"a972e8cd-cae8-4894-b2d3-7bfff75acc1c\") " pod="manila-kuttl-tests/keystone-5f5c86bdc8-jpphg" Feb 02 15:32:55 crc kubenswrapper[4733]: I0202 15:32:55.462155 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a972e8cd-cae8-4894-b2d3-7bfff75acc1c-config-data\") pod \"keystone-5f5c86bdc8-jpphg\" (UID: \"a972e8cd-cae8-4894-b2d3-7bfff75acc1c\") " pod="manila-kuttl-tests/keystone-5f5c86bdc8-jpphg" Feb 02 15:32:55 crc kubenswrapper[4733]: I0202 15:32:55.467875 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a972e8cd-cae8-4894-b2d3-7bfff75acc1c-fernet-keys\") pod \"keystone-5f5c86bdc8-jpphg\" (UID: \"a972e8cd-cae8-4894-b2d3-7bfff75acc1c\") " pod="manila-kuttl-tests/keystone-5f5c86bdc8-jpphg" Feb 02 15:32:55 crc kubenswrapper[4733]: I0202 15:32:55.468775 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a972e8cd-cae8-4894-b2d3-7bfff75acc1c-credential-keys\") pod \"keystone-5f5c86bdc8-jpphg\" (UID: \"a972e8cd-cae8-4894-b2d3-7bfff75acc1c\") " pod="manila-kuttl-tests/keystone-5f5c86bdc8-jpphg" Feb 02 15:32:55 crc kubenswrapper[4733]: I0202 15:32:55.470515 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a972e8cd-cae8-4894-b2d3-7bfff75acc1c-config-data\") pod \"keystone-5f5c86bdc8-jpphg\" (UID: \"a972e8cd-cae8-4894-b2d3-7bfff75acc1c\") " pod="manila-kuttl-tests/keystone-5f5c86bdc8-jpphg" Feb 02 15:32:55 crc kubenswrapper[4733]: I0202 15:32:55.473911 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a972e8cd-cae8-4894-b2d3-7bfff75acc1c-scripts\") pod \"keystone-5f5c86bdc8-jpphg\" (UID: \"a972e8cd-cae8-4894-b2d3-7bfff75acc1c\") " pod="manila-kuttl-tests/keystone-5f5c86bdc8-jpphg" Feb 02 15:32:55 crc kubenswrapper[4733]: I0202 15:32:55.483101 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nr7m\" (UniqueName: \"kubernetes.io/projected/a972e8cd-cae8-4894-b2d3-7bfff75acc1c-kube-api-access-5nr7m\") pod \"keystone-5f5c86bdc8-jpphg\" (UID: \"a972e8cd-cae8-4894-b2d3-7bfff75acc1c\") " pod="manila-kuttl-tests/keystone-5f5c86bdc8-jpphg" Feb 02 15:32:55 crc kubenswrapper[4733]: I0202 15:32:55.543526 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/keystone-5f5c86bdc8-jpphg" Feb 02 15:32:55 crc kubenswrapper[4733]: I0202 15:32:55.952799 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/keystone-5f5c86bdc8-jpphg"] Feb 02 15:32:56 crc kubenswrapper[4733]: I0202 15:32:56.149639 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/keystone-5f5c86bdc8-jpphg" event={"ID":"a972e8cd-cae8-4894-b2d3-7bfff75acc1c","Type":"ContainerStarted","Data":"5da8dbd95ec2753a5a0ea2dcb28b8f2288a6cac18e606e5f795f27e228c125f0"} Feb 02 15:32:57 crc kubenswrapper[4733]: I0202 15:32:57.159968 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/keystone-5f5c86bdc8-jpphg" event={"ID":"a972e8cd-cae8-4894-b2d3-7bfff75acc1c","Type":"ContainerStarted","Data":"818ce15b55c28e765033c326ced7cc8bb1dd7b47fe329b86dc8b9830026fccea"} Feb 02 15:32:57 crc kubenswrapper[4733]: I0202 15:32:57.160599 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="manila-kuttl-tests/keystone-5f5c86bdc8-jpphg" Feb 02 15:32:57 crc kubenswrapper[4733]: I0202 15:32:57.199139 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="manila-kuttl-tests/keystone-5f5c86bdc8-jpphg" podStartSLOduration=2.19911343 podStartE2EDuration="2.19911343s" podCreationTimestamp="2026-02-02 15:32:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:32:57.184336654 +0000 UTC m=+1060.635798042" watchObservedRunningTime="2026-02-02 15:32:57.19911343 +0000 UTC m=+1060.650574798" Feb 02 15:33:04 crc kubenswrapper[4733]: I0202 15:33:04.987636 4733 patch_prober.go:28] interesting pod/machine-config-daemon-zgsv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 15:33:04 crc kubenswrapper[4733]: I0202 15:33:04.988466 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 15:33:04 crc kubenswrapper[4733]: I0202 15:33:04.988538 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" Feb 02 15:33:04 crc kubenswrapper[4733]: I0202 15:33:04.989472 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b050108ccfab6b78f46eeabfccf3d93e55ea50151667d0fea583d0f4073f7662"} pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 15:33:04 crc kubenswrapper[4733]: I0202 15:33:04.989592 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" containerName="machine-config-daemon" containerID="cri-o://b050108ccfab6b78f46eeabfccf3d93e55ea50151667d0fea583d0f4073f7662" gracePeriod=600 Feb 02 15:33:05 crc kubenswrapper[4733]: I0202 15:33:05.436224 4733 generic.go:334] "Generic (PLEG): container finished" podID="83d0e810-ea5c-47aa-947b-7865b045bf94" containerID="b050108ccfab6b78f46eeabfccf3d93e55ea50151667d0fea583d0f4073f7662" exitCode=0 Feb 02 15:33:05 crc kubenswrapper[4733]: I0202 15:33:05.436263 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" event={"ID":"83d0e810-ea5c-47aa-947b-7865b045bf94","Type":"ContainerDied","Data":"b050108ccfab6b78f46eeabfccf3d93e55ea50151667d0fea583d0f4073f7662"} Feb 02 15:33:05 crc kubenswrapper[4733]: I0202 15:33:05.436719 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" event={"ID":"83d0e810-ea5c-47aa-947b-7865b045bf94","Type":"ContainerStarted","Data":"26cec88244c10ffe50680f654ac54834881498bea583f15ab3590dda3791c1bf"} Feb 02 15:33:05 crc kubenswrapper[4733]: I0202 15:33:05.436756 4733 scope.go:117] "RemoveContainer" containerID="25cacf7f1a9153fe0cadb8062fa48df6b9242dcf030528fe627cf68ad889d6b1" Feb 02 15:33:26 crc kubenswrapper[4733]: I0202 15:33:26.948639 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="manila-kuttl-tests/keystone-5f5c86bdc8-jpphg" Feb 02 15:33:33 crc kubenswrapper[4733]: E0202 15:33:33.243111 4733 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.2:39314->38.102.83.2:34939: write tcp 38.102.83.2:39314->38.102.83.2:34939: write: broken pipe Feb 02 15:33:37 crc kubenswrapper[4733]: I0202 15:33:37.337884 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-index-nwnsz"] Feb 02 15:33:37 crc kubenswrapper[4733]: I0202 15:33:37.339100 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-index-nwnsz" Feb 02 15:33:37 crc kubenswrapper[4733]: I0202 15:33:37.342949 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-index-dockercfg-mfhm7" Feb 02 15:33:37 crc kubenswrapper[4733]: I0202 15:33:37.356395 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-index-nwnsz"] Feb 02 15:33:37 crc kubenswrapper[4733]: I0202 15:33:37.448518 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpfzb\" (UniqueName: \"kubernetes.io/projected/bee045d5-602e-4a90-88a8-38f779c2cb2b-kube-api-access-zpfzb\") pod \"manila-operator-index-nwnsz\" (UID: \"bee045d5-602e-4a90-88a8-38f779c2cb2b\") " pod="openstack-operators/manila-operator-index-nwnsz" Feb 02 15:33:37 crc kubenswrapper[4733]: I0202 15:33:37.549481 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpfzb\" (UniqueName: \"kubernetes.io/projected/bee045d5-602e-4a90-88a8-38f779c2cb2b-kube-api-access-zpfzb\") pod \"manila-operator-index-nwnsz\" (UID: \"bee045d5-602e-4a90-88a8-38f779c2cb2b\") " pod="openstack-operators/manila-operator-index-nwnsz" Feb 02 15:33:37 crc kubenswrapper[4733]: I0202 15:33:37.568986 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpfzb\" (UniqueName: \"kubernetes.io/projected/bee045d5-602e-4a90-88a8-38f779c2cb2b-kube-api-access-zpfzb\") pod \"manila-operator-index-nwnsz\" (UID: \"bee045d5-602e-4a90-88a8-38f779c2cb2b\") " pod="openstack-operators/manila-operator-index-nwnsz" Feb 02 15:33:37 crc kubenswrapper[4733]: I0202 15:33:37.659263 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-index-nwnsz" Feb 02 15:33:38 crc kubenswrapper[4733]: I0202 15:33:38.163076 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-index-nwnsz"] Feb 02 15:33:38 crc kubenswrapper[4733]: I0202 15:33:38.698811 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-index-nwnsz" event={"ID":"bee045d5-602e-4a90-88a8-38f779c2cb2b","Type":"ContainerStarted","Data":"3d198ee4094dfab7634ea175b893ff87967c515a0674c0e80b4efdf9c0d6340e"} Feb 02 15:33:41 crc kubenswrapper[4733]: I0202 15:33:41.535120 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/manila-operator-index-nwnsz"] Feb 02 15:33:41 crc kubenswrapper[4733]: I0202 15:33:41.718945 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-index-nwnsz" event={"ID":"bee045d5-602e-4a90-88a8-38f779c2cb2b","Type":"ContainerStarted","Data":"bb5eec46d50d47d20e53146028827370b901d62c6eacdb697341dae105737fbe"} Feb 02 15:33:41 crc kubenswrapper[4733]: I0202 15:33:41.744898 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-index-nwnsz" podStartSLOduration=2.159145588 podStartE2EDuration="4.744880429s" podCreationTimestamp="2026-02-02 15:33:37 +0000 UTC" firstStartedPulling="2026-02-02 15:33:38.174461581 +0000 UTC m=+1101.625922969" lastFinishedPulling="2026-02-02 15:33:40.760196452 +0000 UTC m=+1104.211657810" observedRunningTime="2026-02-02 15:33:41.743359699 +0000 UTC m=+1105.194821047" watchObservedRunningTime="2026-02-02 15:33:41.744880429 +0000 UTC m=+1105.196341787" Feb 02 15:33:42 crc kubenswrapper[4733]: I0202 15:33:42.145526 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-index-gqz7t"] Feb 02 15:33:42 crc kubenswrapper[4733]: I0202 15:33:42.146783 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-index-gqz7t" Feb 02 15:33:42 crc kubenswrapper[4733]: I0202 15:33:42.154639 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-index-gqz7t"] Feb 02 15:33:42 crc kubenswrapper[4733]: I0202 15:33:42.320688 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzgdc\" (UniqueName: \"kubernetes.io/projected/1ae26a29-79a7-4f0a-b546-f20743ba97a3-kube-api-access-fzgdc\") pod \"manila-operator-index-gqz7t\" (UID: \"1ae26a29-79a7-4f0a-b546-f20743ba97a3\") " pod="openstack-operators/manila-operator-index-gqz7t" Feb 02 15:33:42 crc kubenswrapper[4733]: I0202 15:33:42.422021 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzgdc\" (UniqueName: \"kubernetes.io/projected/1ae26a29-79a7-4f0a-b546-f20743ba97a3-kube-api-access-fzgdc\") pod \"manila-operator-index-gqz7t\" (UID: \"1ae26a29-79a7-4f0a-b546-f20743ba97a3\") " pod="openstack-operators/manila-operator-index-gqz7t" Feb 02 15:33:42 crc kubenswrapper[4733]: I0202 15:33:42.444349 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzgdc\" (UniqueName: \"kubernetes.io/projected/1ae26a29-79a7-4f0a-b546-f20743ba97a3-kube-api-access-fzgdc\") pod \"manila-operator-index-gqz7t\" (UID: \"1ae26a29-79a7-4f0a-b546-f20743ba97a3\") " pod="openstack-operators/manila-operator-index-gqz7t" Feb 02 15:33:42 crc kubenswrapper[4733]: I0202 15:33:42.474889 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-index-gqz7t" Feb 02 15:33:42 crc kubenswrapper[4733]: I0202 15:33:42.731582 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/manila-operator-index-nwnsz" podUID="bee045d5-602e-4a90-88a8-38f779c2cb2b" containerName="registry-server" containerID="cri-o://bb5eec46d50d47d20e53146028827370b901d62c6eacdb697341dae105737fbe" gracePeriod=2 Feb 02 15:33:42 crc kubenswrapper[4733]: I0202 15:33:42.762493 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-index-gqz7t"] Feb 02 15:33:42 crc kubenswrapper[4733]: W0202 15:33:42.817970 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1ae26a29_79a7_4f0a_b546_f20743ba97a3.slice/crio-51f3f474f991a56c8b947253da5d206e93197e2fee3103d564ddef69beadf504 WatchSource:0}: Error finding container 51f3f474f991a56c8b947253da5d206e93197e2fee3103d564ddef69beadf504: Status 404 returned error can't find the container with id 51f3f474f991a56c8b947253da5d206e93197e2fee3103d564ddef69beadf504 Feb 02 15:33:43 crc kubenswrapper[4733]: I0202 15:33:43.180330 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-index-nwnsz" Feb 02 15:33:43 crc kubenswrapper[4733]: I0202 15:33:43.335430 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zpfzb\" (UniqueName: \"kubernetes.io/projected/bee045d5-602e-4a90-88a8-38f779c2cb2b-kube-api-access-zpfzb\") pod \"bee045d5-602e-4a90-88a8-38f779c2cb2b\" (UID: \"bee045d5-602e-4a90-88a8-38f779c2cb2b\") " Feb 02 15:33:43 crc kubenswrapper[4733]: I0202 15:33:43.341334 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bee045d5-602e-4a90-88a8-38f779c2cb2b-kube-api-access-zpfzb" (OuterVolumeSpecName: "kube-api-access-zpfzb") pod "bee045d5-602e-4a90-88a8-38f779c2cb2b" (UID: "bee045d5-602e-4a90-88a8-38f779c2cb2b"). InnerVolumeSpecName "kube-api-access-zpfzb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:33:43 crc kubenswrapper[4733]: I0202 15:33:43.437951 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zpfzb\" (UniqueName: \"kubernetes.io/projected/bee045d5-602e-4a90-88a8-38f779c2cb2b-kube-api-access-zpfzb\") on node \"crc\" DevicePath \"\"" Feb 02 15:33:43 crc kubenswrapper[4733]: I0202 15:33:43.741845 4733 generic.go:334] "Generic (PLEG): container finished" podID="bee045d5-602e-4a90-88a8-38f779c2cb2b" containerID="bb5eec46d50d47d20e53146028827370b901d62c6eacdb697341dae105737fbe" exitCode=0 Feb 02 15:33:43 crc kubenswrapper[4733]: I0202 15:33:43.741908 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-index-nwnsz" event={"ID":"bee045d5-602e-4a90-88a8-38f779c2cb2b","Type":"ContainerDied","Data":"bb5eec46d50d47d20e53146028827370b901d62c6eacdb697341dae105737fbe"} Feb 02 15:33:43 crc kubenswrapper[4733]: I0202 15:33:43.741938 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-index-nwnsz" Feb 02 15:33:43 crc kubenswrapper[4733]: I0202 15:33:43.742011 4733 scope.go:117] "RemoveContainer" containerID="bb5eec46d50d47d20e53146028827370b901d62c6eacdb697341dae105737fbe" Feb 02 15:33:43 crc kubenswrapper[4733]: I0202 15:33:43.741990 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-index-nwnsz" event={"ID":"bee045d5-602e-4a90-88a8-38f779c2cb2b","Type":"ContainerDied","Data":"3d198ee4094dfab7634ea175b893ff87967c515a0674c0e80b4efdf9c0d6340e"} Feb 02 15:33:43 crc kubenswrapper[4733]: I0202 15:33:43.744071 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-index-gqz7t" event={"ID":"1ae26a29-79a7-4f0a-b546-f20743ba97a3","Type":"ContainerStarted","Data":"6b98411c753ef28c5622e28730c655cc9b2d9663f5b9c1e89989944c008b5759"} Feb 02 15:33:43 crc kubenswrapper[4733]: I0202 15:33:43.744097 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-index-gqz7t" event={"ID":"1ae26a29-79a7-4f0a-b546-f20743ba97a3","Type":"ContainerStarted","Data":"51f3f474f991a56c8b947253da5d206e93197e2fee3103d564ddef69beadf504"} Feb 02 15:33:43 crc kubenswrapper[4733]: I0202 15:33:43.770955 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-index-gqz7t" podStartSLOduration=1.382520488 podStartE2EDuration="1.77084157s" podCreationTimestamp="2026-02-02 15:33:42 +0000 UTC" firstStartedPulling="2026-02-02 15:33:42.82472083 +0000 UTC m=+1106.276182188" lastFinishedPulling="2026-02-02 15:33:43.213041912 +0000 UTC m=+1106.664503270" observedRunningTime="2026-02-02 15:33:43.764534375 +0000 UTC m=+1107.215995743" watchObservedRunningTime="2026-02-02 15:33:43.77084157 +0000 UTC m=+1107.222302968" Feb 02 15:33:43 crc kubenswrapper[4733]: I0202 15:33:43.780256 4733 scope.go:117] "RemoveContainer" containerID="bb5eec46d50d47d20e53146028827370b901d62c6eacdb697341dae105737fbe" Feb 02 15:33:43 crc kubenswrapper[4733]: E0202 15:33:43.781148 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb5eec46d50d47d20e53146028827370b901d62c6eacdb697341dae105737fbe\": container with ID starting with bb5eec46d50d47d20e53146028827370b901d62c6eacdb697341dae105737fbe not found: ID does not exist" containerID="bb5eec46d50d47d20e53146028827370b901d62c6eacdb697341dae105737fbe" Feb 02 15:33:43 crc kubenswrapper[4733]: I0202 15:33:43.781224 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb5eec46d50d47d20e53146028827370b901d62c6eacdb697341dae105737fbe"} err="failed to get container status \"bb5eec46d50d47d20e53146028827370b901d62c6eacdb697341dae105737fbe\": rpc error: code = NotFound desc = could not find container \"bb5eec46d50d47d20e53146028827370b901d62c6eacdb697341dae105737fbe\": container with ID starting with bb5eec46d50d47d20e53146028827370b901d62c6eacdb697341dae105737fbe not found: ID does not exist" Feb 02 15:33:43 crc kubenswrapper[4733]: I0202 15:33:43.796083 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/manila-operator-index-nwnsz"] Feb 02 15:33:43 crc kubenswrapper[4733]: I0202 15:33:43.806180 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/manila-operator-index-nwnsz"] Feb 02 15:33:45 crc kubenswrapper[4733]: I0202 15:33:45.270314 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bee045d5-602e-4a90-88a8-38f779c2cb2b" path="/var/lib/kubelet/pods/bee045d5-602e-4a90-88a8-38f779c2cb2b/volumes" Feb 02 15:33:52 crc kubenswrapper[4733]: I0202 15:33:52.475744 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/manila-operator-index-gqz7t" Feb 02 15:33:52 crc kubenswrapper[4733]: I0202 15:33:52.476352 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-index-gqz7t" Feb 02 15:33:52 crc kubenswrapper[4733]: I0202 15:33:52.512246 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/manila-operator-index-gqz7t" Feb 02 15:33:52 crc kubenswrapper[4733]: I0202 15:33:52.868048 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-index-gqz7t" Feb 02 15:34:00 crc kubenswrapper[4733]: I0202 15:34:00.987099 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/28815fdbd8da3955636af5641949aa0dc9d10a5ebc7db4178a6eac7a27cmqpd"] Feb 02 15:34:00 crc kubenswrapper[4733]: E0202 15:34:00.988148 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bee045d5-602e-4a90-88a8-38f779c2cb2b" containerName="registry-server" Feb 02 15:34:00 crc kubenswrapper[4733]: I0202 15:34:00.988194 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="bee045d5-602e-4a90-88a8-38f779c2cb2b" containerName="registry-server" Feb 02 15:34:00 crc kubenswrapper[4733]: I0202 15:34:00.988471 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="bee045d5-602e-4a90-88a8-38f779c2cb2b" containerName="registry-server" Feb 02 15:34:00 crc kubenswrapper[4733]: I0202 15:34:00.990339 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/28815fdbd8da3955636af5641949aa0dc9d10a5ebc7db4178a6eac7a27cmqpd" Feb 02 15:34:00 crc kubenswrapper[4733]: I0202 15:34:00.992544 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-2mtp8" Feb 02 15:34:00 crc kubenswrapper[4733]: I0202 15:34:00.993660 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/28815fdbd8da3955636af5641949aa0dc9d10a5ebc7db4178a6eac7a27cmqpd"] Feb 02 15:34:01 crc kubenswrapper[4733]: I0202 15:34:01.095277 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9-bundle\") pod \"28815fdbd8da3955636af5641949aa0dc9d10a5ebc7db4178a6eac7a27cmqpd\" (UID: \"56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9\") " pod="openstack-operators/28815fdbd8da3955636af5641949aa0dc9d10a5ebc7db4178a6eac7a27cmqpd" Feb 02 15:34:01 crc kubenswrapper[4733]: I0202 15:34:01.095366 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2j24n\" (UniqueName: \"kubernetes.io/projected/56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9-kube-api-access-2j24n\") pod \"28815fdbd8da3955636af5641949aa0dc9d10a5ebc7db4178a6eac7a27cmqpd\" (UID: \"56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9\") " pod="openstack-operators/28815fdbd8da3955636af5641949aa0dc9d10a5ebc7db4178a6eac7a27cmqpd" Feb 02 15:34:01 crc kubenswrapper[4733]: I0202 15:34:01.095406 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9-util\") pod \"28815fdbd8da3955636af5641949aa0dc9d10a5ebc7db4178a6eac7a27cmqpd\" (UID: \"56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9\") " pod="openstack-operators/28815fdbd8da3955636af5641949aa0dc9d10a5ebc7db4178a6eac7a27cmqpd" Feb 02 15:34:01 crc kubenswrapper[4733]: I0202 15:34:01.196482 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9-bundle\") pod \"28815fdbd8da3955636af5641949aa0dc9d10a5ebc7db4178a6eac7a27cmqpd\" (UID: \"56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9\") " pod="openstack-operators/28815fdbd8da3955636af5641949aa0dc9d10a5ebc7db4178a6eac7a27cmqpd" Feb 02 15:34:01 crc kubenswrapper[4733]: I0202 15:34:01.196572 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2j24n\" (UniqueName: \"kubernetes.io/projected/56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9-kube-api-access-2j24n\") pod \"28815fdbd8da3955636af5641949aa0dc9d10a5ebc7db4178a6eac7a27cmqpd\" (UID: \"56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9\") " pod="openstack-operators/28815fdbd8da3955636af5641949aa0dc9d10a5ebc7db4178a6eac7a27cmqpd" Feb 02 15:34:01 crc kubenswrapper[4733]: I0202 15:34:01.196604 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9-util\") pod \"28815fdbd8da3955636af5641949aa0dc9d10a5ebc7db4178a6eac7a27cmqpd\" (UID: \"56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9\") " pod="openstack-operators/28815fdbd8da3955636af5641949aa0dc9d10a5ebc7db4178a6eac7a27cmqpd" Feb 02 15:34:01 crc kubenswrapper[4733]: I0202 15:34:01.197218 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9-util\") pod \"28815fdbd8da3955636af5641949aa0dc9d10a5ebc7db4178a6eac7a27cmqpd\" (UID: \"56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9\") " pod="openstack-operators/28815fdbd8da3955636af5641949aa0dc9d10a5ebc7db4178a6eac7a27cmqpd" Feb 02 15:34:01 crc kubenswrapper[4733]: I0202 15:34:01.197482 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9-bundle\") pod \"28815fdbd8da3955636af5641949aa0dc9d10a5ebc7db4178a6eac7a27cmqpd\" (UID: \"56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9\") " pod="openstack-operators/28815fdbd8da3955636af5641949aa0dc9d10a5ebc7db4178a6eac7a27cmqpd" Feb 02 15:34:01 crc kubenswrapper[4733]: I0202 15:34:01.224981 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2j24n\" (UniqueName: \"kubernetes.io/projected/56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9-kube-api-access-2j24n\") pod \"28815fdbd8da3955636af5641949aa0dc9d10a5ebc7db4178a6eac7a27cmqpd\" (UID: \"56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9\") " pod="openstack-operators/28815fdbd8da3955636af5641949aa0dc9d10a5ebc7db4178a6eac7a27cmqpd" Feb 02 15:34:01 crc kubenswrapper[4733]: I0202 15:34:01.314828 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/28815fdbd8da3955636af5641949aa0dc9d10a5ebc7db4178a6eac7a27cmqpd" Feb 02 15:34:01 crc kubenswrapper[4733]: I0202 15:34:01.751741 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/28815fdbd8da3955636af5641949aa0dc9d10a5ebc7db4178a6eac7a27cmqpd"] Feb 02 15:34:01 crc kubenswrapper[4733]: W0202 15:34:01.762423 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod56e9f0c8_ecc5_4af6_87c6_7cbeab57e5f9.slice/crio-4fba9c6df789b6a00a17f485e7ce7bea4a2f7beef47a40b643c80d8dc0ee130a WatchSource:0}: Error finding container 4fba9c6df789b6a00a17f485e7ce7bea4a2f7beef47a40b643c80d8dc0ee130a: Status 404 returned error can't find the container with id 4fba9c6df789b6a00a17f485e7ce7bea4a2f7beef47a40b643c80d8dc0ee130a Feb 02 15:34:01 crc kubenswrapper[4733]: I0202 15:34:01.898415 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/28815fdbd8da3955636af5641949aa0dc9d10a5ebc7db4178a6eac7a27cmqpd" event={"ID":"56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9","Type":"ContainerStarted","Data":"4fba9c6df789b6a00a17f485e7ce7bea4a2f7beef47a40b643c80d8dc0ee130a"} Feb 02 15:34:02 crc kubenswrapper[4733]: I0202 15:34:02.907027 4733 generic.go:334] "Generic (PLEG): container finished" podID="56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9" containerID="c8112f53db575592095d0a3ac1b3f517d2cd02c636461f258aad664c4709316f" exitCode=0 Feb 02 15:34:02 crc kubenswrapper[4733]: I0202 15:34:02.907332 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/28815fdbd8da3955636af5641949aa0dc9d10a5ebc7db4178a6eac7a27cmqpd" event={"ID":"56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9","Type":"ContainerDied","Data":"c8112f53db575592095d0a3ac1b3f517d2cd02c636461f258aad664c4709316f"} Feb 02 15:34:03 crc kubenswrapper[4733]: I0202 15:34:03.916415 4733 generic.go:334] "Generic (PLEG): container finished" podID="56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9" containerID="bdbb55dce4f250f0293be7e174c78454be12f47e5dd826641c4dd8a005f5bf3f" exitCode=0 Feb 02 15:34:03 crc kubenswrapper[4733]: I0202 15:34:03.916496 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/28815fdbd8da3955636af5641949aa0dc9d10a5ebc7db4178a6eac7a27cmqpd" event={"ID":"56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9","Type":"ContainerDied","Data":"bdbb55dce4f250f0293be7e174c78454be12f47e5dd826641c4dd8a005f5bf3f"} Feb 02 15:34:04 crc kubenswrapper[4733]: I0202 15:34:04.926088 4733 generic.go:334] "Generic (PLEG): container finished" podID="56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9" containerID="d4696ce0ed2dab71f0282caae5dd7bafb4b5e9a7b706a558a2cca872fa7edde3" exitCode=0 Feb 02 15:34:04 crc kubenswrapper[4733]: I0202 15:34:04.926204 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/28815fdbd8da3955636af5641949aa0dc9d10a5ebc7db4178a6eac7a27cmqpd" event={"ID":"56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9","Type":"ContainerDied","Data":"d4696ce0ed2dab71f0282caae5dd7bafb4b5e9a7b706a558a2cca872fa7edde3"} Feb 02 15:34:06 crc kubenswrapper[4733]: I0202 15:34:06.224540 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/28815fdbd8da3955636af5641949aa0dc9d10a5ebc7db4178a6eac7a27cmqpd" Feb 02 15:34:06 crc kubenswrapper[4733]: I0202 15:34:06.370400 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9-util\") pod \"56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9\" (UID: \"56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9\") " Feb 02 15:34:06 crc kubenswrapper[4733]: I0202 15:34:06.370488 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9-bundle\") pod \"56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9\" (UID: \"56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9\") " Feb 02 15:34:06 crc kubenswrapper[4733]: I0202 15:34:06.370609 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2j24n\" (UniqueName: \"kubernetes.io/projected/56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9-kube-api-access-2j24n\") pod \"56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9\" (UID: \"56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9\") " Feb 02 15:34:06 crc kubenswrapper[4733]: I0202 15:34:06.372641 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9-bundle" (OuterVolumeSpecName: "bundle") pod "56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9" (UID: "56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:34:06 crc kubenswrapper[4733]: I0202 15:34:06.376242 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9-kube-api-access-2j24n" (OuterVolumeSpecName: "kube-api-access-2j24n") pod "56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9" (UID: "56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9"). InnerVolumeSpecName "kube-api-access-2j24n". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:34:06 crc kubenswrapper[4733]: I0202 15:34:06.400211 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9-util" (OuterVolumeSpecName: "util") pod "56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9" (UID: "56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:34:06 crc kubenswrapper[4733]: I0202 15:34:06.472414 4733 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 15:34:06 crc kubenswrapper[4733]: I0202 15:34:06.472462 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2j24n\" (UniqueName: \"kubernetes.io/projected/56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9-kube-api-access-2j24n\") on node \"crc\" DevicePath \"\"" Feb 02 15:34:06 crc kubenswrapper[4733]: I0202 15:34:06.472477 4733 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9-util\") on node \"crc\" DevicePath \"\"" Feb 02 15:34:06 crc kubenswrapper[4733]: I0202 15:34:06.947975 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/28815fdbd8da3955636af5641949aa0dc9d10a5ebc7db4178a6eac7a27cmqpd" event={"ID":"56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9","Type":"ContainerDied","Data":"4fba9c6df789b6a00a17f485e7ce7bea4a2f7beef47a40b643c80d8dc0ee130a"} Feb 02 15:34:06 crc kubenswrapper[4733]: I0202 15:34:06.948039 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4fba9c6df789b6a00a17f485e7ce7bea4a2f7beef47a40b643c80d8dc0ee130a" Feb 02 15:34:06 crc kubenswrapper[4733]: I0202 15:34:06.948066 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/28815fdbd8da3955636af5641949aa0dc9d10a5ebc7db4178a6eac7a27cmqpd" Feb 02 15:34:17 crc kubenswrapper[4733]: I0202 15:34:17.186246 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-64f9574466-56vg6"] Feb 02 15:34:17 crc kubenswrapper[4733]: E0202 15:34:17.187437 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9" containerName="pull" Feb 02 15:34:17 crc kubenswrapper[4733]: I0202 15:34:17.187453 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9" containerName="pull" Feb 02 15:34:17 crc kubenswrapper[4733]: E0202 15:34:17.187465 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9" containerName="util" Feb 02 15:34:17 crc kubenswrapper[4733]: I0202 15:34:17.187473 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9" containerName="util" Feb 02 15:34:17 crc kubenswrapper[4733]: E0202 15:34:17.187489 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9" containerName="extract" Feb 02 15:34:17 crc kubenswrapper[4733]: I0202 15:34:17.187496 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9" containerName="extract" Feb 02 15:34:17 crc kubenswrapper[4733]: I0202 15:34:17.187628 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9" containerName="extract" Feb 02 15:34:17 crc kubenswrapper[4733]: I0202 15:34:17.188143 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-64f9574466-56vg6" Feb 02 15:34:17 crc kubenswrapper[4733]: I0202 15:34:17.190365 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-8b856" Feb 02 15:34:17 crc kubenswrapper[4733]: I0202 15:34:17.190517 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-service-cert" Feb 02 15:34:17 crc kubenswrapper[4733]: I0202 15:34:17.327644 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-64f9574466-56vg6"] Feb 02 15:34:17 crc kubenswrapper[4733]: I0202 15:34:17.331537 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f3971f39-e65f-4262-acaf-ca5fca3f7cc9-webhook-cert\") pod \"manila-operator-controller-manager-64f9574466-56vg6\" (UID: \"f3971f39-e65f-4262-acaf-ca5fca3f7cc9\") " pod="openstack-operators/manila-operator-controller-manager-64f9574466-56vg6" Feb 02 15:34:17 crc kubenswrapper[4733]: I0202 15:34:17.331649 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f3971f39-e65f-4262-acaf-ca5fca3f7cc9-apiservice-cert\") pod \"manila-operator-controller-manager-64f9574466-56vg6\" (UID: \"f3971f39-e65f-4262-acaf-ca5fca3f7cc9\") " pod="openstack-operators/manila-operator-controller-manager-64f9574466-56vg6" Feb 02 15:34:17 crc kubenswrapper[4733]: I0202 15:34:17.331685 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzrxk\" (UniqueName: \"kubernetes.io/projected/f3971f39-e65f-4262-acaf-ca5fca3f7cc9-kube-api-access-mzrxk\") pod \"manila-operator-controller-manager-64f9574466-56vg6\" (UID: \"f3971f39-e65f-4262-acaf-ca5fca3f7cc9\") " pod="openstack-operators/manila-operator-controller-manager-64f9574466-56vg6" Feb 02 15:34:17 crc kubenswrapper[4733]: I0202 15:34:17.433144 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f3971f39-e65f-4262-acaf-ca5fca3f7cc9-apiservice-cert\") pod \"manila-operator-controller-manager-64f9574466-56vg6\" (UID: \"f3971f39-e65f-4262-acaf-ca5fca3f7cc9\") " pod="openstack-operators/manila-operator-controller-manager-64f9574466-56vg6" Feb 02 15:34:17 crc kubenswrapper[4733]: I0202 15:34:17.433231 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzrxk\" (UniqueName: \"kubernetes.io/projected/f3971f39-e65f-4262-acaf-ca5fca3f7cc9-kube-api-access-mzrxk\") pod \"manila-operator-controller-manager-64f9574466-56vg6\" (UID: \"f3971f39-e65f-4262-acaf-ca5fca3f7cc9\") " pod="openstack-operators/manila-operator-controller-manager-64f9574466-56vg6" Feb 02 15:34:17 crc kubenswrapper[4733]: I0202 15:34:17.433254 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f3971f39-e65f-4262-acaf-ca5fca3f7cc9-webhook-cert\") pod \"manila-operator-controller-manager-64f9574466-56vg6\" (UID: \"f3971f39-e65f-4262-acaf-ca5fca3f7cc9\") " pod="openstack-operators/manila-operator-controller-manager-64f9574466-56vg6" Feb 02 15:34:17 crc kubenswrapper[4733]: I0202 15:34:17.438517 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f3971f39-e65f-4262-acaf-ca5fca3f7cc9-webhook-cert\") pod \"manila-operator-controller-manager-64f9574466-56vg6\" (UID: \"f3971f39-e65f-4262-acaf-ca5fca3f7cc9\") " pod="openstack-operators/manila-operator-controller-manager-64f9574466-56vg6" Feb 02 15:34:17 crc kubenswrapper[4733]: I0202 15:34:17.438571 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f3971f39-e65f-4262-acaf-ca5fca3f7cc9-apiservice-cert\") pod \"manila-operator-controller-manager-64f9574466-56vg6\" (UID: \"f3971f39-e65f-4262-acaf-ca5fca3f7cc9\") " pod="openstack-operators/manila-operator-controller-manager-64f9574466-56vg6" Feb 02 15:34:17 crc kubenswrapper[4733]: I0202 15:34:17.449534 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzrxk\" (UniqueName: \"kubernetes.io/projected/f3971f39-e65f-4262-acaf-ca5fca3f7cc9-kube-api-access-mzrxk\") pod \"manila-operator-controller-manager-64f9574466-56vg6\" (UID: \"f3971f39-e65f-4262-acaf-ca5fca3f7cc9\") " pod="openstack-operators/manila-operator-controller-manager-64f9574466-56vg6" Feb 02 15:34:17 crc kubenswrapper[4733]: I0202 15:34:17.508833 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-64f9574466-56vg6" Feb 02 15:34:17 crc kubenswrapper[4733]: I0202 15:34:17.925093 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-64f9574466-56vg6"] Feb 02 15:34:17 crc kubenswrapper[4733]: W0202 15:34:17.934071 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf3971f39_e65f_4262_acaf_ca5fca3f7cc9.slice/crio-de176ba5e529933724fa6a4378d07fb016b609f3778e811313906fcbf5442dfd WatchSource:0}: Error finding container de176ba5e529933724fa6a4378d07fb016b609f3778e811313906fcbf5442dfd: Status 404 returned error can't find the container with id de176ba5e529933724fa6a4378d07fb016b609f3778e811313906fcbf5442dfd Feb 02 15:34:18 crc kubenswrapper[4733]: I0202 15:34:18.030017 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-64f9574466-56vg6" event={"ID":"f3971f39-e65f-4262-acaf-ca5fca3f7cc9","Type":"ContainerStarted","Data":"de176ba5e529933724fa6a4378d07fb016b609f3778e811313906fcbf5442dfd"} Feb 02 15:34:20 crc kubenswrapper[4733]: I0202 15:34:20.042102 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-64f9574466-56vg6" event={"ID":"f3971f39-e65f-4262-acaf-ca5fca3f7cc9","Type":"ContainerStarted","Data":"84e34713078767c3675090908ff02de8038baca92063e381c481b5e26d015c5b"} Feb 02 15:34:20 crc kubenswrapper[4733]: I0202 15:34:20.066410 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-64f9574466-56vg6" podStartSLOduration=1.815679377 podStartE2EDuration="3.066390611s" podCreationTimestamp="2026-02-02 15:34:17 +0000 UTC" firstStartedPulling="2026-02-02 15:34:17.938019615 +0000 UTC m=+1141.389480973" lastFinishedPulling="2026-02-02 15:34:19.188730849 +0000 UTC m=+1142.640192207" observedRunningTime="2026-02-02 15:34:20.060195549 +0000 UTC m=+1143.511656927" watchObservedRunningTime="2026-02-02 15:34:20.066390611 +0000 UTC m=+1143.517851969" Feb 02 15:34:21 crc kubenswrapper[4733]: I0202 15:34:21.048218 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-64f9574466-56vg6" Feb 02 15:34:27 crc kubenswrapper[4733]: I0202 15:34:27.513127 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-64f9574466-56vg6" Feb 02 15:34:31 crc kubenswrapper[4733]: I0202 15:34:31.400186 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["manila-kuttl-tests/manila-db-create-zlfrz"] Feb 02 15:34:31 crc kubenswrapper[4733]: I0202 15:34:31.400948 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-db-create-zlfrz" Feb 02 15:34:31 crc kubenswrapper[4733]: I0202 15:34:31.421006 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["manila-kuttl-tests/manila-6c72-account-create-update-fkts6"] Feb 02 15:34:31 crc kubenswrapper[4733]: I0202 15:34:31.422569 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-6c72-account-create-update-fkts6" Feb 02 15:34:31 crc kubenswrapper[4733]: I0202 15:34:31.431185 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"manila-db-secret" Feb 02 15:34:31 crc kubenswrapper[4733]: I0202 15:34:31.438462 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/manila-db-create-zlfrz"] Feb 02 15:34:31 crc kubenswrapper[4733]: I0202 15:34:31.464017 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/manila-6c72-account-create-update-fkts6"] Feb 02 15:34:31 crc kubenswrapper[4733]: I0202 15:34:31.532705 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhgvs\" (UniqueName: \"kubernetes.io/projected/ae455916-7075-477b-a71f-04537587c8ac-kube-api-access-qhgvs\") pod \"manila-db-create-zlfrz\" (UID: \"ae455916-7075-477b-a71f-04537587c8ac\") " pod="manila-kuttl-tests/manila-db-create-zlfrz" Feb 02 15:34:31 crc kubenswrapper[4733]: I0202 15:34:31.532761 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae455916-7075-477b-a71f-04537587c8ac-operator-scripts\") pod \"manila-db-create-zlfrz\" (UID: \"ae455916-7075-477b-a71f-04537587c8ac\") " pod="manila-kuttl-tests/manila-db-create-zlfrz" Feb 02 15:34:31 crc kubenswrapper[4733]: I0202 15:34:31.532793 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhj2s\" (UniqueName: \"kubernetes.io/projected/77b06765-7874-4e7c-a681-d5da283dae25-kube-api-access-dhj2s\") pod \"manila-6c72-account-create-update-fkts6\" (UID: \"77b06765-7874-4e7c-a681-d5da283dae25\") " pod="manila-kuttl-tests/manila-6c72-account-create-update-fkts6" Feb 02 15:34:31 crc kubenswrapper[4733]: I0202 15:34:31.532873 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77b06765-7874-4e7c-a681-d5da283dae25-operator-scripts\") pod \"manila-6c72-account-create-update-fkts6\" (UID: \"77b06765-7874-4e7c-a681-d5da283dae25\") " pod="manila-kuttl-tests/manila-6c72-account-create-update-fkts6" Feb 02 15:34:31 crc kubenswrapper[4733]: I0202 15:34:31.633931 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhgvs\" (UniqueName: \"kubernetes.io/projected/ae455916-7075-477b-a71f-04537587c8ac-kube-api-access-qhgvs\") pod \"manila-db-create-zlfrz\" (UID: \"ae455916-7075-477b-a71f-04537587c8ac\") " pod="manila-kuttl-tests/manila-db-create-zlfrz" Feb 02 15:34:31 crc kubenswrapper[4733]: I0202 15:34:31.633986 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae455916-7075-477b-a71f-04537587c8ac-operator-scripts\") pod \"manila-db-create-zlfrz\" (UID: \"ae455916-7075-477b-a71f-04537587c8ac\") " pod="manila-kuttl-tests/manila-db-create-zlfrz" Feb 02 15:34:31 crc kubenswrapper[4733]: I0202 15:34:31.634011 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhj2s\" (UniqueName: \"kubernetes.io/projected/77b06765-7874-4e7c-a681-d5da283dae25-kube-api-access-dhj2s\") pod \"manila-6c72-account-create-update-fkts6\" (UID: \"77b06765-7874-4e7c-a681-d5da283dae25\") " pod="manila-kuttl-tests/manila-6c72-account-create-update-fkts6" Feb 02 15:34:31 crc kubenswrapper[4733]: I0202 15:34:31.634052 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77b06765-7874-4e7c-a681-d5da283dae25-operator-scripts\") pod \"manila-6c72-account-create-update-fkts6\" (UID: \"77b06765-7874-4e7c-a681-d5da283dae25\") " pod="manila-kuttl-tests/manila-6c72-account-create-update-fkts6" Feb 02 15:34:31 crc kubenswrapper[4733]: I0202 15:34:31.635028 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77b06765-7874-4e7c-a681-d5da283dae25-operator-scripts\") pod \"manila-6c72-account-create-update-fkts6\" (UID: \"77b06765-7874-4e7c-a681-d5da283dae25\") " pod="manila-kuttl-tests/manila-6c72-account-create-update-fkts6" Feb 02 15:34:31 crc kubenswrapper[4733]: I0202 15:34:31.635227 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae455916-7075-477b-a71f-04537587c8ac-operator-scripts\") pod \"manila-db-create-zlfrz\" (UID: \"ae455916-7075-477b-a71f-04537587c8ac\") " pod="manila-kuttl-tests/manila-db-create-zlfrz" Feb 02 15:34:31 crc kubenswrapper[4733]: I0202 15:34:31.659094 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhgvs\" (UniqueName: \"kubernetes.io/projected/ae455916-7075-477b-a71f-04537587c8ac-kube-api-access-qhgvs\") pod \"manila-db-create-zlfrz\" (UID: \"ae455916-7075-477b-a71f-04537587c8ac\") " pod="manila-kuttl-tests/manila-db-create-zlfrz" Feb 02 15:34:31 crc kubenswrapper[4733]: I0202 15:34:31.664993 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhj2s\" (UniqueName: \"kubernetes.io/projected/77b06765-7874-4e7c-a681-d5da283dae25-kube-api-access-dhj2s\") pod \"manila-6c72-account-create-update-fkts6\" (UID: \"77b06765-7874-4e7c-a681-d5da283dae25\") " pod="manila-kuttl-tests/manila-6c72-account-create-update-fkts6" Feb 02 15:34:31 crc kubenswrapper[4733]: I0202 15:34:31.720010 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-db-create-zlfrz" Feb 02 15:34:31 crc kubenswrapper[4733]: I0202 15:34:31.774001 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-6c72-account-create-update-fkts6" Feb 02 15:34:32 crc kubenswrapper[4733]: I0202 15:34:32.199901 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/manila-db-create-zlfrz"] Feb 02 15:34:32 crc kubenswrapper[4733]: I0202 15:34:32.304658 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/manila-6c72-account-create-update-fkts6"] Feb 02 15:34:32 crc kubenswrapper[4733]: W0202 15:34:32.306414 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod77b06765_7874_4e7c_a681_d5da283dae25.slice/crio-3880a9f6fba4b78746a757b578a68d1dec5719af0937b26739aafd3e483e8c9f WatchSource:0}: Error finding container 3880a9f6fba4b78746a757b578a68d1dec5719af0937b26739aafd3e483e8c9f: Status 404 returned error can't find the container with id 3880a9f6fba4b78746a757b578a68d1dec5719af0937b26739aafd3e483e8c9f Feb 02 15:34:33 crc kubenswrapper[4733]: I0202 15:34:33.136905 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-6c72-account-create-update-fkts6" event={"ID":"77b06765-7874-4e7c-a681-d5da283dae25","Type":"ContainerStarted","Data":"ad074bd93f775844cad04e7d9ac407c52dca41251c0106fba43a6c42de4a1e7e"} Feb 02 15:34:33 crc kubenswrapper[4733]: I0202 15:34:33.136952 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-6c72-account-create-update-fkts6" event={"ID":"77b06765-7874-4e7c-a681-d5da283dae25","Type":"ContainerStarted","Data":"3880a9f6fba4b78746a757b578a68d1dec5719af0937b26739aafd3e483e8c9f"} Feb 02 15:34:33 crc kubenswrapper[4733]: I0202 15:34:33.145577 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-db-create-zlfrz" event={"ID":"ae455916-7075-477b-a71f-04537587c8ac","Type":"ContainerStarted","Data":"09989f03d83fa6e83049f56a1c6e1ae9819f524cb9c149681d71df8727e1096e"} Feb 02 15:34:33 crc kubenswrapper[4733]: I0202 15:34:33.145620 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-db-create-zlfrz" event={"ID":"ae455916-7075-477b-a71f-04537587c8ac","Type":"ContainerStarted","Data":"3c08878c8a2dfb76d68671a1b6a18e3c7a4682df2a29d21a7c80d7d7343c9e31"} Feb 02 15:34:33 crc kubenswrapper[4733]: I0202 15:34:33.158749 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="manila-kuttl-tests/manila-6c72-account-create-update-fkts6" podStartSLOduration=2.15873172 podStartE2EDuration="2.15873172s" podCreationTimestamp="2026-02-02 15:34:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:34:33.153914174 +0000 UTC m=+1156.605375532" watchObservedRunningTime="2026-02-02 15:34:33.15873172 +0000 UTC m=+1156.610193078" Feb 02 15:34:33 crc kubenswrapper[4733]: I0202 15:34:33.183838 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="manila-kuttl-tests/manila-db-create-zlfrz" podStartSLOduration=2.183812835 podStartE2EDuration="2.183812835s" podCreationTimestamp="2026-02-02 15:34:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:34:33.177337846 +0000 UTC m=+1156.628799204" watchObservedRunningTime="2026-02-02 15:34:33.183812835 +0000 UTC m=+1156.635274203" Feb 02 15:34:34 crc kubenswrapper[4733]: I0202 15:34:34.156225 4733 generic.go:334] "Generic (PLEG): container finished" podID="ae455916-7075-477b-a71f-04537587c8ac" containerID="09989f03d83fa6e83049f56a1c6e1ae9819f524cb9c149681d71df8727e1096e" exitCode=0 Feb 02 15:34:34 crc kubenswrapper[4733]: I0202 15:34:34.156274 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-db-create-zlfrz" event={"ID":"ae455916-7075-477b-a71f-04537587c8ac","Type":"ContainerDied","Data":"09989f03d83fa6e83049f56a1c6e1ae9819f524cb9c149681d71df8727e1096e"} Feb 02 15:34:34 crc kubenswrapper[4733]: I0202 15:34:34.158712 4733 generic.go:334] "Generic (PLEG): container finished" podID="77b06765-7874-4e7c-a681-d5da283dae25" containerID="ad074bd93f775844cad04e7d9ac407c52dca41251c0106fba43a6c42de4a1e7e" exitCode=0 Feb 02 15:34:34 crc kubenswrapper[4733]: I0202 15:34:34.158773 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-6c72-account-create-update-fkts6" event={"ID":"77b06765-7874-4e7c-a681-d5da283dae25","Type":"ContainerDied","Data":"ad074bd93f775844cad04e7d9ac407c52dca41251c0106fba43a6c42de4a1e7e"} Feb 02 15:34:35 crc kubenswrapper[4733]: I0202 15:34:35.657154 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-db-create-zlfrz" Feb 02 15:34:35 crc kubenswrapper[4733]: I0202 15:34:35.661380 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-6c72-account-create-update-fkts6" Feb 02 15:34:35 crc kubenswrapper[4733]: I0202 15:34:35.822255 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dhj2s\" (UniqueName: \"kubernetes.io/projected/77b06765-7874-4e7c-a681-d5da283dae25-kube-api-access-dhj2s\") pod \"77b06765-7874-4e7c-a681-d5da283dae25\" (UID: \"77b06765-7874-4e7c-a681-d5da283dae25\") " Feb 02 15:34:35 crc kubenswrapper[4733]: I0202 15:34:35.822350 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qhgvs\" (UniqueName: \"kubernetes.io/projected/ae455916-7075-477b-a71f-04537587c8ac-kube-api-access-qhgvs\") pod \"ae455916-7075-477b-a71f-04537587c8ac\" (UID: \"ae455916-7075-477b-a71f-04537587c8ac\") " Feb 02 15:34:35 crc kubenswrapper[4733]: I0202 15:34:35.822400 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77b06765-7874-4e7c-a681-d5da283dae25-operator-scripts\") pod \"77b06765-7874-4e7c-a681-d5da283dae25\" (UID: \"77b06765-7874-4e7c-a681-d5da283dae25\") " Feb 02 15:34:35 crc kubenswrapper[4733]: I0202 15:34:35.822495 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae455916-7075-477b-a71f-04537587c8ac-operator-scripts\") pod \"ae455916-7075-477b-a71f-04537587c8ac\" (UID: \"ae455916-7075-477b-a71f-04537587c8ac\") " Feb 02 15:34:35 crc kubenswrapper[4733]: I0202 15:34:35.822925 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae455916-7075-477b-a71f-04537587c8ac-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ae455916-7075-477b-a71f-04537587c8ac" (UID: "ae455916-7075-477b-a71f-04537587c8ac"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:34:35 crc kubenswrapper[4733]: I0202 15:34:35.823247 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77b06765-7874-4e7c-a681-d5da283dae25-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "77b06765-7874-4e7c-a681-d5da283dae25" (UID: "77b06765-7874-4e7c-a681-d5da283dae25"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:34:35 crc kubenswrapper[4733]: I0202 15:34:35.830770 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77b06765-7874-4e7c-a681-d5da283dae25-kube-api-access-dhj2s" (OuterVolumeSpecName: "kube-api-access-dhj2s") pod "77b06765-7874-4e7c-a681-d5da283dae25" (UID: "77b06765-7874-4e7c-a681-d5da283dae25"). InnerVolumeSpecName "kube-api-access-dhj2s". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:34:35 crc kubenswrapper[4733]: I0202 15:34:35.830969 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae455916-7075-477b-a71f-04537587c8ac-kube-api-access-qhgvs" (OuterVolumeSpecName: "kube-api-access-qhgvs") pod "ae455916-7075-477b-a71f-04537587c8ac" (UID: "ae455916-7075-477b-a71f-04537587c8ac"). InnerVolumeSpecName "kube-api-access-qhgvs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:34:35 crc kubenswrapper[4733]: I0202 15:34:35.923559 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qhgvs\" (UniqueName: \"kubernetes.io/projected/ae455916-7075-477b-a71f-04537587c8ac-kube-api-access-qhgvs\") on node \"crc\" DevicePath \"\"" Feb 02 15:34:35 crc kubenswrapper[4733]: I0202 15:34:35.923888 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77b06765-7874-4e7c-a681-d5da283dae25-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 15:34:35 crc kubenswrapper[4733]: I0202 15:34:35.923995 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae455916-7075-477b-a71f-04537587c8ac-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 15:34:35 crc kubenswrapper[4733]: I0202 15:34:35.924071 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dhj2s\" (UniqueName: \"kubernetes.io/projected/77b06765-7874-4e7c-a681-d5da283dae25-kube-api-access-dhj2s\") on node \"crc\" DevicePath \"\"" Feb 02 15:34:36 crc kubenswrapper[4733]: I0202 15:34:36.177232 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-db-create-zlfrz" event={"ID":"ae455916-7075-477b-a71f-04537587c8ac","Type":"ContainerDied","Data":"3c08878c8a2dfb76d68671a1b6a18e3c7a4682df2a29d21a7c80d7d7343c9e31"} Feb 02 15:34:36 crc kubenswrapper[4733]: I0202 15:34:36.177529 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3c08878c8a2dfb76d68671a1b6a18e3c7a4682df2a29d21a7c80d7d7343c9e31" Feb 02 15:34:36 crc kubenswrapper[4733]: I0202 15:34:36.177299 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-db-create-zlfrz" Feb 02 15:34:36 crc kubenswrapper[4733]: I0202 15:34:36.178971 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-6c72-account-create-update-fkts6" event={"ID":"77b06765-7874-4e7c-a681-d5da283dae25","Type":"ContainerDied","Data":"3880a9f6fba4b78746a757b578a68d1dec5719af0937b26739aafd3e483e8c9f"} Feb 02 15:34:36 crc kubenswrapper[4733]: I0202 15:34:36.179013 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3880a9f6fba4b78746a757b578a68d1dec5719af0937b26739aafd3e483e8c9f" Feb 02 15:34:36 crc kubenswrapper[4733]: I0202 15:34:36.179059 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-6c72-account-create-update-fkts6" Feb 02 15:34:41 crc kubenswrapper[4733]: I0202 15:34:41.652033 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["manila-kuttl-tests/manila-db-sync-fr7nc"] Feb 02 15:34:41 crc kubenswrapper[4733]: E0202 15:34:41.652705 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae455916-7075-477b-a71f-04537587c8ac" containerName="mariadb-database-create" Feb 02 15:34:41 crc kubenswrapper[4733]: I0202 15:34:41.652724 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae455916-7075-477b-a71f-04537587c8ac" containerName="mariadb-database-create" Feb 02 15:34:41 crc kubenswrapper[4733]: E0202 15:34:41.652754 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77b06765-7874-4e7c-a681-d5da283dae25" containerName="mariadb-account-create-update" Feb 02 15:34:41 crc kubenswrapper[4733]: I0202 15:34:41.652765 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="77b06765-7874-4e7c-a681-d5da283dae25" containerName="mariadb-account-create-update" Feb 02 15:34:41 crc kubenswrapper[4733]: I0202 15:34:41.652906 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae455916-7075-477b-a71f-04537587c8ac" containerName="mariadb-database-create" Feb 02 15:34:41 crc kubenswrapper[4733]: I0202 15:34:41.652921 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="77b06765-7874-4e7c-a681-d5da283dae25" containerName="mariadb-account-create-update" Feb 02 15:34:41 crc kubenswrapper[4733]: I0202 15:34:41.653467 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-db-sync-fr7nc" Feb 02 15:34:41 crc kubenswrapper[4733]: I0202 15:34:41.656730 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"manila-config-data" Feb 02 15:34:41 crc kubenswrapper[4733]: I0202 15:34:41.658626 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"manila-manila-dockercfg-mm6xc" Feb 02 15:34:41 crc kubenswrapper[4733]: I0202 15:34:41.662539 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/manila-db-sync-fr7nc"] Feb 02 15:34:41 crc kubenswrapper[4733]: I0202 15:34:41.815154 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/eb872024-d057-4eaa-8f93-0401ac466650-job-config-data\") pod \"manila-db-sync-fr7nc\" (UID: \"eb872024-d057-4eaa-8f93-0401ac466650\") " pod="manila-kuttl-tests/manila-db-sync-fr7nc" Feb 02 15:34:41 crc kubenswrapper[4733]: I0202 15:34:41.815591 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77lpn\" (UniqueName: \"kubernetes.io/projected/eb872024-d057-4eaa-8f93-0401ac466650-kube-api-access-77lpn\") pod \"manila-db-sync-fr7nc\" (UID: \"eb872024-d057-4eaa-8f93-0401ac466650\") " pod="manila-kuttl-tests/manila-db-sync-fr7nc" Feb 02 15:34:41 crc kubenswrapper[4733]: I0202 15:34:41.815681 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb872024-d057-4eaa-8f93-0401ac466650-config-data\") pod \"manila-db-sync-fr7nc\" (UID: \"eb872024-d057-4eaa-8f93-0401ac466650\") " pod="manila-kuttl-tests/manila-db-sync-fr7nc" Feb 02 15:34:41 crc kubenswrapper[4733]: I0202 15:34:41.917380 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77lpn\" (UniqueName: \"kubernetes.io/projected/eb872024-d057-4eaa-8f93-0401ac466650-kube-api-access-77lpn\") pod \"manila-db-sync-fr7nc\" (UID: \"eb872024-d057-4eaa-8f93-0401ac466650\") " pod="manila-kuttl-tests/manila-db-sync-fr7nc" Feb 02 15:34:41 crc kubenswrapper[4733]: I0202 15:34:41.917519 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb872024-d057-4eaa-8f93-0401ac466650-config-data\") pod \"manila-db-sync-fr7nc\" (UID: \"eb872024-d057-4eaa-8f93-0401ac466650\") " pod="manila-kuttl-tests/manila-db-sync-fr7nc" Feb 02 15:34:41 crc kubenswrapper[4733]: I0202 15:34:41.917556 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/eb872024-d057-4eaa-8f93-0401ac466650-job-config-data\") pod \"manila-db-sync-fr7nc\" (UID: \"eb872024-d057-4eaa-8f93-0401ac466650\") " pod="manila-kuttl-tests/manila-db-sync-fr7nc" Feb 02 15:34:41 crc kubenswrapper[4733]: I0202 15:34:41.924659 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb872024-d057-4eaa-8f93-0401ac466650-config-data\") pod \"manila-db-sync-fr7nc\" (UID: \"eb872024-d057-4eaa-8f93-0401ac466650\") " pod="manila-kuttl-tests/manila-db-sync-fr7nc" Feb 02 15:34:41 crc kubenswrapper[4733]: I0202 15:34:41.925013 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/eb872024-d057-4eaa-8f93-0401ac466650-job-config-data\") pod \"manila-db-sync-fr7nc\" (UID: \"eb872024-d057-4eaa-8f93-0401ac466650\") " pod="manila-kuttl-tests/manila-db-sync-fr7nc" Feb 02 15:34:41 crc kubenswrapper[4733]: I0202 15:34:41.952359 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77lpn\" (UniqueName: \"kubernetes.io/projected/eb872024-d057-4eaa-8f93-0401ac466650-kube-api-access-77lpn\") pod \"manila-db-sync-fr7nc\" (UID: \"eb872024-d057-4eaa-8f93-0401ac466650\") " pod="manila-kuttl-tests/manila-db-sync-fr7nc" Feb 02 15:34:41 crc kubenswrapper[4733]: I0202 15:34:41.976649 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-db-sync-fr7nc" Feb 02 15:34:42 crc kubenswrapper[4733]: I0202 15:34:42.439986 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/manila-db-sync-fr7nc"] Feb 02 15:34:42 crc kubenswrapper[4733]: W0202 15:34:42.448720 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb872024_d057_4eaa_8f93_0401ac466650.slice/crio-e7c98d0d9121b850d64f63a4e44228a5f0a5a690b7218cf4313aa36dc6bdb555 WatchSource:0}: Error finding container e7c98d0d9121b850d64f63a4e44228a5f0a5a690b7218cf4313aa36dc6bdb555: Status 404 returned error can't find the container with id e7c98d0d9121b850d64f63a4e44228a5f0a5a690b7218cf4313aa36dc6bdb555 Feb 02 15:34:43 crc kubenswrapper[4733]: I0202 15:34:43.242001 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-db-sync-fr7nc" event={"ID":"eb872024-d057-4eaa-8f93-0401ac466650","Type":"ContainerStarted","Data":"e7c98d0d9121b850d64f63a4e44228a5f0a5a690b7218cf4313aa36dc6bdb555"} Feb 02 15:34:47 crc kubenswrapper[4733]: I0202 15:34:47.276260 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-db-sync-fr7nc" event={"ID":"eb872024-d057-4eaa-8f93-0401ac466650","Type":"ContainerStarted","Data":"16745d57b94a753d9e93ffa48cc9c02bda6166e8d95e1a988a017658b15ed912"} Feb 02 15:34:47 crc kubenswrapper[4733]: I0202 15:34:47.314192 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="manila-kuttl-tests/manila-db-sync-fr7nc" podStartSLOduration=2.236097244 podStartE2EDuration="6.314154116s" podCreationTimestamp="2026-02-02 15:34:41 +0000 UTC" firstStartedPulling="2026-02-02 15:34:42.452922295 +0000 UTC m=+1165.904383703" lastFinishedPulling="2026-02-02 15:34:46.530979217 +0000 UTC m=+1169.982440575" observedRunningTime="2026-02-02 15:34:47.310092099 +0000 UTC m=+1170.761553477" watchObservedRunningTime="2026-02-02 15:34:47.314154116 +0000 UTC m=+1170.765615464" Feb 02 15:35:08 crc kubenswrapper[4733]: I0202 15:35:08.481007 4733 generic.go:334] "Generic (PLEG): container finished" podID="eb872024-d057-4eaa-8f93-0401ac466650" containerID="16745d57b94a753d9e93ffa48cc9c02bda6166e8d95e1a988a017658b15ed912" exitCode=0 Feb 02 15:35:08 crc kubenswrapper[4733]: I0202 15:35:08.481146 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-db-sync-fr7nc" event={"ID":"eb872024-d057-4eaa-8f93-0401ac466650","Type":"ContainerDied","Data":"16745d57b94a753d9e93ffa48cc9c02bda6166e8d95e1a988a017658b15ed912"} Feb 02 15:35:09 crc kubenswrapper[4733]: I0202 15:35:09.909368 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-db-sync-fr7nc" Feb 02 15:35:09 crc kubenswrapper[4733]: I0202 15:35:09.952685 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb872024-d057-4eaa-8f93-0401ac466650-config-data\") pod \"eb872024-d057-4eaa-8f93-0401ac466650\" (UID: \"eb872024-d057-4eaa-8f93-0401ac466650\") " Feb 02 15:35:09 crc kubenswrapper[4733]: I0202 15:35:09.952840 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-77lpn\" (UniqueName: \"kubernetes.io/projected/eb872024-d057-4eaa-8f93-0401ac466650-kube-api-access-77lpn\") pod \"eb872024-d057-4eaa-8f93-0401ac466650\" (UID: \"eb872024-d057-4eaa-8f93-0401ac466650\") " Feb 02 15:35:09 crc kubenswrapper[4733]: I0202 15:35:09.952872 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/eb872024-d057-4eaa-8f93-0401ac466650-job-config-data\") pod \"eb872024-d057-4eaa-8f93-0401ac466650\" (UID: \"eb872024-d057-4eaa-8f93-0401ac466650\") " Feb 02 15:35:09 crc kubenswrapper[4733]: I0202 15:35:09.966735 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb872024-d057-4eaa-8f93-0401ac466650-kube-api-access-77lpn" (OuterVolumeSpecName: "kube-api-access-77lpn") pod "eb872024-d057-4eaa-8f93-0401ac466650" (UID: "eb872024-d057-4eaa-8f93-0401ac466650"). InnerVolumeSpecName "kube-api-access-77lpn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:35:09 crc kubenswrapper[4733]: I0202 15:35:09.966899 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb872024-d057-4eaa-8f93-0401ac466650-job-config-data" (OuterVolumeSpecName: "job-config-data") pod "eb872024-d057-4eaa-8f93-0401ac466650" (UID: "eb872024-d057-4eaa-8f93-0401ac466650"). InnerVolumeSpecName "job-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:35:09 crc kubenswrapper[4733]: I0202 15:35:09.970511 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb872024-d057-4eaa-8f93-0401ac466650-config-data" (OuterVolumeSpecName: "config-data") pod "eb872024-d057-4eaa-8f93-0401ac466650" (UID: "eb872024-d057-4eaa-8f93-0401ac466650"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:35:10 crc kubenswrapper[4733]: I0202 15:35:10.054098 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb872024-d057-4eaa-8f93-0401ac466650-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 15:35:10 crc kubenswrapper[4733]: I0202 15:35:10.054141 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-77lpn\" (UniqueName: \"kubernetes.io/projected/eb872024-d057-4eaa-8f93-0401ac466650-kube-api-access-77lpn\") on node \"crc\" DevicePath \"\"" Feb 02 15:35:10 crc kubenswrapper[4733]: I0202 15:35:10.054157 4733 reconciler_common.go:293] "Volume detached for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/eb872024-d057-4eaa-8f93-0401ac466650-job-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 15:35:10 crc kubenswrapper[4733]: I0202 15:35:10.502147 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-db-sync-fr7nc" event={"ID":"eb872024-d057-4eaa-8f93-0401ac466650","Type":"ContainerDied","Data":"e7c98d0d9121b850d64f63a4e44228a5f0a5a690b7218cf4313aa36dc6bdb555"} Feb 02 15:35:10 crc kubenswrapper[4733]: I0202 15:35:10.502254 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e7c98d0d9121b850d64f63a4e44228a5f0a5a690b7218cf4313aa36dc6bdb555" Feb 02 15:35:10 crc kubenswrapper[4733]: I0202 15:35:10.502333 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-db-sync-fr7nc" Feb 02 15:35:10 crc kubenswrapper[4733]: I0202 15:35:10.875604 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["manila-kuttl-tests/manila-scheduler-0"] Feb 02 15:35:10 crc kubenswrapper[4733]: E0202 15:35:10.876088 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb872024-d057-4eaa-8f93-0401ac466650" containerName="manila-db-sync" Feb 02 15:35:10 crc kubenswrapper[4733]: I0202 15:35:10.876121 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb872024-d057-4eaa-8f93-0401ac466650" containerName="manila-db-sync" Feb 02 15:35:10 crc kubenswrapper[4733]: I0202 15:35:10.876364 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb872024-d057-4eaa-8f93-0401ac466650" containerName="manila-db-sync" Feb 02 15:35:10 crc kubenswrapper[4733]: I0202 15:35:10.877585 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:35:10 crc kubenswrapper[4733]: I0202 15:35:10.884454 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"manila-manila-dockercfg-mm6xc" Feb 02 15:35:10 crc kubenswrapper[4733]: I0202 15:35:10.884644 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"manila-scripts" Feb 02 15:35:10 crc kubenswrapper[4733]: I0202 15:35:10.884651 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"manila-config-data" Feb 02 15:35:10 crc kubenswrapper[4733]: I0202 15:35:10.884782 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"manila-scheduler-config-data" Feb 02 15:35:10 crc kubenswrapper[4733]: I0202 15:35:10.889119 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/manila-scheduler-0"] Feb 02 15:35:10 crc kubenswrapper[4733]: I0202 15:35:10.904886 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["manila-kuttl-tests/manila-share-share0-0"] Feb 02 15:35:10 crc kubenswrapper[4733]: I0202 15:35:10.906183 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:35:10 crc kubenswrapper[4733]: I0202 15:35:10.908316 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"ceph-conf-files" Feb 02 15:35:10 crc kubenswrapper[4733]: I0202 15:35:10.908525 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"manila-share-share0-config-data" Feb 02 15:35:10 crc kubenswrapper[4733]: I0202 15:35:10.936464 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/manila-share-share0-0"] Feb 02 15:35:10 crc kubenswrapper[4733]: I0202 15:35:10.968270 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-th7nw\" (UniqueName: \"kubernetes.io/projected/c5853503-348a-4386-9fec-cfa5b11b7501-kube-api-access-th7nw\") pod \"manila-share-share0-0\" (UID: \"c5853503-348a-4386-9fec-cfa5b11b7501\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:35:10 crc kubenswrapper[4733]: I0202 15:35:10.968313 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c5853503-348a-4386-9fec-cfa5b11b7501-ceph\") pod \"manila-share-share0-0\" (UID: \"c5853503-348a-4386-9fec-cfa5b11b7501\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:35:10 crc kubenswrapper[4733]: I0202 15:35:10.968333 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c5853503-348a-4386-9fec-cfa5b11b7501-etc-machine-id\") pod \"manila-share-share0-0\" (UID: \"c5853503-348a-4386-9fec-cfa5b11b7501\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:35:10 crc kubenswrapper[4733]: I0202 15:35:10.968350 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5853503-348a-4386-9fec-cfa5b11b7501-config-data\") pod \"manila-share-share0-0\" (UID: \"c5853503-348a-4386-9fec-cfa5b11b7501\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:35:10 crc kubenswrapper[4733]: I0202 15:35:10.968482 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drzqn\" (UniqueName: \"kubernetes.io/projected/d04d46e8-50b1-47ab-ac62-a7f446949411-kube-api-access-drzqn\") pod \"manila-scheduler-0\" (UID: \"d04d46e8-50b1-47ab-ac62-a7f446949411\") " pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:35:10 crc kubenswrapper[4733]: I0202 15:35:10.968524 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d04d46e8-50b1-47ab-ac62-a7f446949411-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"d04d46e8-50b1-47ab-ac62-a7f446949411\") " pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:35:10 crc kubenswrapper[4733]: I0202 15:35:10.968622 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d04d46e8-50b1-47ab-ac62-a7f446949411-config-data\") pod \"manila-scheduler-0\" (UID: \"d04d46e8-50b1-47ab-ac62-a7f446949411\") " pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:35:10 crc kubenswrapper[4733]: I0202 15:35:10.968640 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d04d46e8-50b1-47ab-ac62-a7f446949411-scripts\") pod \"manila-scheduler-0\" (UID: \"d04d46e8-50b1-47ab-ac62-a7f446949411\") " pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:35:10 crc kubenswrapper[4733]: I0202 15:35:10.968671 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5853503-348a-4386-9fec-cfa5b11b7501-scripts\") pod \"manila-share-share0-0\" (UID: \"c5853503-348a-4386-9fec-cfa5b11b7501\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:35:10 crc kubenswrapper[4733]: I0202 15:35:10.968693 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c5853503-348a-4386-9fec-cfa5b11b7501-config-data-custom\") pod \"manila-share-share0-0\" (UID: \"c5853503-348a-4386-9fec-cfa5b11b7501\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:35:10 crc kubenswrapper[4733]: I0202 15:35:10.968713 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d04d46e8-50b1-47ab-ac62-a7f446949411-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"d04d46e8-50b1-47ab-ac62-a7f446949411\") " pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:35:10 crc kubenswrapper[4733]: I0202 15:35:10.968731 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/c5853503-348a-4386-9fec-cfa5b11b7501-var-lib-manila\") pod \"manila-share-share0-0\" (UID: \"c5853503-348a-4386-9fec-cfa5b11b7501\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:35:10 crc kubenswrapper[4733]: I0202 15:35:10.999477 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["manila-kuttl-tests/manila-api-0"] Feb 02 15:35:11 crc kubenswrapper[4733]: I0202 15:35:11.000848 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-api-0" Feb 02 15:35:11 crc kubenswrapper[4733]: I0202 15:35:11.003203 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"manila-api-config-data" Feb 02 15:35:11 crc kubenswrapper[4733]: I0202 15:35:11.029148 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/manila-api-0"] Feb 02 15:35:11 crc kubenswrapper[4733]: I0202 15:35:11.070369 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drzqn\" (UniqueName: \"kubernetes.io/projected/d04d46e8-50b1-47ab-ac62-a7f446949411-kube-api-access-drzqn\") pod \"manila-scheduler-0\" (UID: \"d04d46e8-50b1-47ab-ac62-a7f446949411\") " pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:35:11 crc kubenswrapper[4733]: I0202 15:35:11.070422 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d04d46e8-50b1-47ab-ac62-a7f446949411-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"d04d46e8-50b1-47ab-ac62-a7f446949411\") " pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:35:11 crc kubenswrapper[4733]: I0202 15:35:11.070469 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d04d46e8-50b1-47ab-ac62-a7f446949411-config-data\") pod \"manila-scheduler-0\" (UID: \"d04d46e8-50b1-47ab-ac62-a7f446949411\") " pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:35:11 crc kubenswrapper[4733]: I0202 15:35:11.070486 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d04d46e8-50b1-47ab-ac62-a7f446949411-scripts\") pod \"manila-scheduler-0\" (UID: \"d04d46e8-50b1-47ab-ac62-a7f446949411\") " pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:35:11 crc kubenswrapper[4733]: I0202 15:35:11.070506 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5853503-348a-4386-9fec-cfa5b11b7501-scripts\") pod \"manila-share-share0-0\" (UID: \"c5853503-348a-4386-9fec-cfa5b11b7501\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:35:11 crc kubenswrapper[4733]: I0202 15:35:11.070520 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c5853503-348a-4386-9fec-cfa5b11b7501-config-data-custom\") pod \"manila-share-share0-0\" (UID: \"c5853503-348a-4386-9fec-cfa5b11b7501\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:35:11 crc kubenswrapper[4733]: I0202 15:35:11.070538 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d04d46e8-50b1-47ab-ac62-a7f446949411-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"d04d46e8-50b1-47ab-ac62-a7f446949411\") " pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:35:11 crc kubenswrapper[4733]: I0202 15:35:11.070552 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/c5853503-348a-4386-9fec-cfa5b11b7501-var-lib-manila\") pod \"manila-share-share0-0\" (UID: \"c5853503-348a-4386-9fec-cfa5b11b7501\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:35:11 crc kubenswrapper[4733]: I0202 15:35:11.070574 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a-logs\") pod \"manila-api-0\" (UID: \"b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:35:11 crc kubenswrapper[4733]: I0202 15:35:11.070589 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4d6d\" (UniqueName: \"kubernetes.io/projected/b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a-kube-api-access-g4d6d\") pod \"manila-api-0\" (UID: \"b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:35:11 crc kubenswrapper[4733]: I0202 15:35:11.070623 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a-config-data\") pod \"manila-api-0\" (UID: \"b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:35:11 crc kubenswrapper[4733]: I0202 15:35:11.070643 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-th7nw\" (UniqueName: \"kubernetes.io/projected/c5853503-348a-4386-9fec-cfa5b11b7501-kube-api-access-th7nw\") pod \"manila-share-share0-0\" (UID: \"c5853503-348a-4386-9fec-cfa5b11b7501\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:35:11 crc kubenswrapper[4733]: I0202 15:35:11.070660 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c5853503-348a-4386-9fec-cfa5b11b7501-ceph\") pod \"manila-share-share0-0\" (UID: \"c5853503-348a-4386-9fec-cfa5b11b7501\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:35:11 crc kubenswrapper[4733]: I0202 15:35:11.070676 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c5853503-348a-4386-9fec-cfa5b11b7501-etc-machine-id\") pod \"manila-share-share0-0\" (UID: \"c5853503-348a-4386-9fec-cfa5b11b7501\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:35:11 crc kubenswrapper[4733]: I0202 15:35:11.070691 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5853503-348a-4386-9fec-cfa5b11b7501-config-data\") pod \"manila-share-share0-0\" (UID: \"c5853503-348a-4386-9fec-cfa5b11b7501\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:35:11 crc kubenswrapper[4733]: I0202 15:35:11.070705 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a-config-data-custom\") pod \"manila-api-0\" (UID: \"b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:35:11 crc kubenswrapper[4733]: I0202 15:35:11.070719 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a-scripts\") pod \"manila-api-0\" (UID: \"b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:35:11 crc kubenswrapper[4733]: I0202 15:35:11.070745 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a-etc-machine-id\") pod \"manila-api-0\" (UID: \"b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:35:11 crc kubenswrapper[4733]: I0202 15:35:11.070817 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d04d46e8-50b1-47ab-ac62-a7f446949411-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"d04d46e8-50b1-47ab-ac62-a7f446949411\") " pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:35:11 crc kubenswrapper[4733]: I0202 15:35:11.071899 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/c5853503-348a-4386-9fec-cfa5b11b7501-var-lib-manila\") pod \"manila-share-share0-0\" (UID: \"c5853503-348a-4386-9fec-cfa5b11b7501\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:35:11 crc kubenswrapper[4733]: I0202 15:35:11.071971 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c5853503-348a-4386-9fec-cfa5b11b7501-etc-machine-id\") pod \"manila-share-share0-0\" (UID: \"c5853503-348a-4386-9fec-cfa5b11b7501\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:35:11 crc kubenswrapper[4733]: I0202 15:35:11.073978 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d04d46e8-50b1-47ab-ac62-a7f446949411-scripts\") pod \"manila-scheduler-0\" (UID: \"d04d46e8-50b1-47ab-ac62-a7f446949411\") " pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:35:11 crc kubenswrapper[4733]: I0202 15:35:11.074317 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d04d46e8-50b1-47ab-ac62-a7f446949411-config-data\") pod \"manila-scheduler-0\" (UID: \"d04d46e8-50b1-47ab-ac62-a7f446949411\") " pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:35:11 crc kubenswrapper[4733]: I0202 15:35:11.074722 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5853503-348a-4386-9fec-cfa5b11b7501-config-data\") pod \"manila-share-share0-0\" (UID: \"c5853503-348a-4386-9fec-cfa5b11b7501\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:35:11 crc kubenswrapper[4733]: I0202 15:35:11.075681 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c5853503-348a-4386-9fec-cfa5b11b7501-ceph\") pod \"manila-share-share0-0\" (UID: \"c5853503-348a-4386-9fec-cfa5b11b7501\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:35:11 crc kubenswrapper[4733]: I0202 15:35:11.076286 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5853503-348a-4386-9fec-cfa5b11b7501-scripts\") pod \"manila-share-share0-0\" (UID: \"c5853503-348a-4386-9fec-cfa5b11b7501\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:35:11 crc kubenswrapper[4733]: I0202 15:35:11.076504 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d04d46e8-50b1-47ab-ac62-a7f446949411-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"d04d46e8-50b1-47ab-ac62-a7f446949411\") " pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:35:11 crc kubenswrapper[4733]: I0202 15:35:11.077370 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c5853503-348a-4386-9fec-cfa5b11b7501-config-data-custom\") pod \"manila-share-share0-0\" (UID: \"c5853503-348a-4386-9fec-cfa5b11b7501\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:35:11 crc kubenswrapper[4733]: I0202 15:35:11.088314 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drzqn\" (UniqueName: \"kubernetes.io/projected/d04d46e8-50b1-47ab-ac62-a7f446949411-kube-api-access-drzqn\") pod \"manila-scheduler-0\" (UID: \"d04d46e8-50b1-47ab-ac62-a7f446949411\") " pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:35:11 crc kubenswrapper[4733]: I0202 15:35:11.088463 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-th7nw\" (UniqueName: \"kubernetes.io/projected/c5853503-348a-4386-9fec-cfa5b11b7501-kube-api-access-th7nw\") pod \"manila-share-share0-0\" (UID: \"c5853503-348a-4386-9fec-cfa5b11b7501\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:35:11 crc kubenswrapper[4733]: I0202 15:35:11.171587 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a-config-data-custom\") pod \"manila-api-0\" (UID: \"b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:35:11 crc kubenswrapper[4733]: I0202 15:35:11.171634 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a-scripts\") pod \"manila-api-0\" (UID: \"b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:35:11 crc kubenswrapper[4733]: I0202 15:35:11.171677 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a-etc-machine-id\") pod \"manila-api-0\" (UID: \"b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:35:11 crc kubenswrapper[4733]: I0202 15:35:11.171749 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a-logs\") pod \"manila-api-0\" (UID: \"b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:35:11 crc kubenswrapper[4733]: I0202 15:35:11.171771 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4d6d\" (UniqueName: \"kubernetes.io/projected/b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a-kube-api-access-g4d6d\") pod \"manila-api-0\" (UID: \"b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:35:11 crc kubenswrapper[4733]: I0202 15:35:11.171809 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a-config-data\") pod \"manila-api-0\" (UID: \"b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:35:11 crc kubenswrapper[4733]: I0202 15:35:11.172034 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a-etc-machine-id\") pod \"manila-api-0\" (UID: \"b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:35:11 crc kubenswrapper[4733]: I0202 15:35:11.172823 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a-logs\") pod \"manila-api-0\" (UID: \"b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:35:11 crc kubenswrapper[4733]: I0202 15:35:11.175206 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a-config-data\") pod \"manila-api-0\" (UID: \"b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:35:11 crc kubenswrapper[4733]: I0202 15:35:11.175626 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a-scripts\") pod \"manila-api-0\" (UID: \"b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:35:11 crc kubenswrapper[4733]: I0202 15:35:11.176339 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a-config-data-custom\") pod \"manila-api-0\" (UID: \"b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:35:11 crc kubenswrapper[4733]: I0202 15:35:11.187462 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4d6d\" (UniqueName: \"kubernetes.io/projected/b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a-kube-api-access-g4d6d\") pod \"manila-api-0\" (UID: \"b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:35:11 crc kubenswrapper[4733]: I0202 15:35:11.196449 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:35:11 crc kubenswrapper[4733]: I0202 15:35:11.224296 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:35:11 crc kubenswrapper[4733]: I0202 15:35:11.315866 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-api-0" Feb 02 15:35:11 crc kubenswrapper[4733]: I0202 15:35:11.516689 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/manila-scheduler-0"] Feb 02 15:35:11 crc kubenswrapper[4733]: I0202 15:35:11.572717 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/manila-api-0"] Feb 02 15:35:11 crc kubenswrapper[4733]: I0202 15:35:11.647991 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/manila-share-share0-0"] Feb 02 15:35:12 crc kubenswrapper[4733]: I0202 15:35:12.516739 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-share-share0-0" event={"ID":"c5853503-348a-4386-9fec-cfa5b11b7501","Type":"ContainerStarted","Data":"c7b4d5086bca48301da0fee5eb1dac5a6b53ed5f8e8d8d2838325c643016ab25"} Feb 02 15:35:12 crc kubenswrapper[4733]: I0202 15:35:12.518601 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-api-0" event={"ID":"b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a","Type":"ContainerStarted","Data":"a68fb69fc8aa87f439ba650174df8155e78d902fa293b457e31c20c2112e9bb7"} Feb 02 15:35:12 crc kubenswrapper[4733]: I0202 15:35:12.518622 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-api-0" event={"ID":"b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a","Type":"ContainerStarted","Data":"a180a6fc32410e33617fc26a50c345f446bcc4a648a07a9f2d7b522c877aa0d1"} Feb 02 15:35:12 crc kubenswrapper[4733]: I0202 15:35:12.518632 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-api-0" event={"ID":"b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a","Type":"ContainerStarted","Data":"2db7ec4764867ecfd65a0119375eb4ad8b7a57046d012cf74333eb00a046f8c5"} Feb 02 15:35:12 crc kubenswrapper[4733]: I0202 15:35:12.518733 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="manila-kuttl-tests/manila-api-0" Feb 02 15:35:12 crc kubenswrapper[4733]: I0202 15:35:12.519917 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-scheduler-0" event={"ID":"d04d46e8-50b1-47ab-ac62-a7f446949411","Type":"ContainerStarted","Data":"1bc04e3fe0d9ec9a8cf631d5103d6ae3b217ccd6ae514fe3fe4634171563531b"} Feb 02 15:35:12 crc kubenswrapper[4733]: I0202 15:35:12.541972 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="manila-kuttl-tests/manila-api-0" podStartSLOduration=2.541956013 podStartE2EDuration="2.541956013s" podCreationTimestamp="2026-02-02 15:35:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:35:12.536492379 +0000 UTC m=+1195.987953737" watchObservedRunningTime="2026-02-02 15:35:12.541956013 +0000 UTC m=+1195.993417371" Feb 02 15:35:13 crc kubenswrapper[4733]: I0202 15:35:13.529664 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-scheduler-0" event={"ID":"d04d46e8-50b1-47ab-ac62-a7f446949411","Type":"ContainerStarted","Data":"be664a164e3540eacbc8f61c59d2f24b886ed58c44ac7a162eb05192733017a5"} Feb 02 15:35:13 crc kubenswrapper[4733]: I0202 15:35:13.530003 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-scheduler-0" event={"ID":"d04d46e8-50b1-47ab-ac62-a7f446949411","Type":"ContainerStarted","Data":"df75409bdb9d1900e091e8eac282b55cecc19bc761f50fd523a5af5a3b567d13"} Feb 02 15:35:13 crc kubenswrapper[4733]: I0202 15:35:13.561027 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="manila-kuttl-tests/manila-scheduler-0" podStartSLOduration=2.827660025 podStartE2EDuration="3.561008437s" podCreationTimestamp="2026-02-02 15:35:10 +0000 UTC" firstStartedPulling="2026-02-02 15:35:11.523573368 +0000 UTC m=+1194.975034736" lastFinishedPulling="2026-02-02 15:35:12.25692179 +0000 UTC m=+1195.708383148" observedRunningTime="2026-02-02 15:35:13.552695797 +0000 UTC m=+1197.004157155" watchObservedRunningTime="2026-02-02 15:35:13.561008437 +0000 UTC m=+1197.012469785" Feb 02 15:35:17 crc kubenswrapper[4733]: I0202 15:35:17.562801 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-share-share0-0" event={"ID":"c5853503-348a-4386-9fec-cfa5b11b7501","Type":"ContainerStarted","Data":"c8d9156f304d30a3ef84f3b663ebb0f580e2927d42d46e4cde55497b6111d850"} Feb 02 15:35:18 crc kubenswrapper[4733]: I0202 15:35:18.575093 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-share-share0-0" event={"ID":"c5853503-348a-4386-9fec-cfa5b11b7501","Type":"ContainerStarted","Data":"116efcfe9abd31a050f2d88c205ff735cb117b043fea831b68e74f6b6fc08b1f"} Feb 02 15:35:18 crc kubenswrapper[4733]: I0202 15:35:18.601627 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="manila-kuttl-tests/manila-share-share0-0" podStartSLOduration=3.410861959 podStartE2EDuration="8.60160396s" podCreationTimestamp="2026-02-02 15:35:10 +0000 UTC" firstStartedPulling="2026-02-02 15:35:11.658808542 +0000 UTC m=+1195.110269900" lastFinishedPulling="2026-02-02 15:35:16.849550543 +0000 UTC m=+1200.301011901" observedRunningTime="2026-02-02 15:35:18.597664976 +0000 UTC m=+1202.049126344" watchObservedRunningTime="2026-02-02 15:35:18.60160396 +0000 UTC m=+1202.053065328" Feb 02 15:35:21 crc kubenswrapper[4733]: I0202 15:35:21.196939 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:35:21 crc kubenswrapper[4733]: I0202 15:35:21.224662 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:35:32 crc kubenswrapper[4733]: I0202 15:35:32.623199 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:35:32 crc kubenswrapper[4733]: I0202 15:35:32.705659 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:35:32 crc kubenswrapper[4733]: I0202 15:35:32.747938 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="manila-kuttl-tests/manila-api-0" Feb 02 15:35:34 crc kubenswrapper[4733]: I0202 15:35:34.395059 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["manila-kuttl-tests/manila-api-2"] Feb 02 15:35:34 crc kubenswrapper[4733]: I0202 15:35:34.396472 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-api-2" Feb 02 15:35:34 crc kubenswrapper[4733]: I0202 15:35:34.414122 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["manila-kuttl-tests/manila-api-1"] Feb 02 15:35:34 crc kubenswrapper[4733]: I0202 15:35:34.415892 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-api-1" Feb 02 15:35:34 crc kubenswrapper[4733]: I0202 15:35:34.433794 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/manila-api-2"] Feb 02 15:35:34 crc kubenswrapper[4733]: I0202 15:35:34.441041 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/manila-api-1"] Feb 02 15:35:34 crc kubenswrapper[4733]: I0202 15:35:34.466533 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e2b1c4e-2735-437e-b2a1-482c2f909d8f-config-data\") pod \"manila-api-1\" (UID: \"2e2b1c4e-2735-437e-b2a1-482c2f909d8f\") " pod="manila-kuttl-tests/manila-api-1" Feb 02 15:35:34 crc kubenswrapper[4733]: I0202 15:35:34.466583 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b9b45f1-f282-4118-b627-94e4f8167d00-logs\") pod \"manila-api-2\" (UID: \"0b9b45f1-f282-4118-b627-94e4f8167d00\") " pod="manila-kuttl-tests/manila-api-2" Feb 02 15:35:34 crc kubenswrapper[4733]: I0202 15:35:34.466607 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e2b1c4e-2735-437e-b2a1-482c2f909d8f-scripts\") pod \"manila-api-1\" (UID: \"2e2b1c4e-2735-437e-b2a1-482c2f909d8f\") " pod="manila-kuttl-tests/manila-api-1" Feb 02 15:35:34 crc kubenswrapper[4733]: I0202 15:35:34.466622 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2e2b1c4e-2735-437e-b2a1-482c2f909d8f-config-data-custom\") pod \"manila-api-1\" (UID: \"2e2b1c4e-2735-437e-b2a1-482c2f909d8f\") " pod="manila-kuttl-tests/manila-api-1" Feb 02 15:35:34 crc kubenswrapper[4733]: I0202 15:35:34.466692 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2e2b1c4e-2735-437e-b2a1-482c2f909d8f-etc-machine-id\") pod \"manila-api-1\" (UID: \"2e2b1c4e-2735-437e-b2a1-482c2f909d8f\") " pod="manila-kuttl-tests/manila-api-1" Feb 02 15:35:34 crc kubenswrapper[4733]: I0202 15:35:34.466735 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkkfp\" (UniqueName: \"kubernetes.io/projected/0b9b45f1-f282-4118-b627-94e4f8167d00-kube-api-access-nkkfp\") pod \"manila-api-2\" (UID: \"0b9b45f1-f282-4118-b627-94e4f8167d00\") " pod="manila-kuttl-tests/manila-api-2" Feb 02 15:35:34 crc kubenswrapper[4733]: I0202 15:35:34.466802 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e2b1c4e-2735-437e-b2a1-482c2f909d8f-logs\") pod \"manila-api-1\" (UID: \"2e2b1c4e-2735-437e-b2a1-482c2f909d8f\") " pod="manila-kuttl-tests/manila-api-1" Feb 02 15:35:34 crc kubenswrapper[4733]: I0202 15:35:34.466836 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0b9b45f1-f282-4118-b627-94e4f8167d00-config-data-custom\") pod \"manila-api-2\" (UID: \"0b9b45f1-f282-4118-b627-94e4f8167d00\") " pod="manila-kuttl-tests/manila-api-2" Feb 02 15:35:34 crc kubenswrapper[4733]: I0202 15:35:34.466862 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b9b45f1-f282-4118-b627-94e4f8167d00-config-data\") pod \"manila-api-2\" (UID: \"0b9b45f1-f282-4118-b627-94e4f8167d00\") " pod="manila-kuttl-tests/manila-api-2" Feb 02 15:35:34 crc kubenswrapper[4733]: I0202 15:35:34.466888 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b9b45f1-f282-4118-b627-94e4f8167d00-scripts\") pod \"manila-api-2\" (UID: \"0b9b45f1-f282-4118-b627-94e4f8167d00\") " pod="manila-kuttl-tests/manila-api-2" Feb 02 15:35:34 crc kubenswrapper[4733]: I0202 15:35:34.466922 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5wkc\" (UniqueName: \"kubernetes.io/projected/2e2b1c4e-2735-437e-b2a1-482c2f909d8f-kube-api-access-x5wkc\") pod \"manila-api-1\" (UID: \"2e2b1c4e-2735-437e-b2a1-482c2f909d8f\") " pod="manila-kuttl-tests/manila-api-1" Feb 02 15:35:34 crc kubenswrapper[4733]: I0202 15:35:34.466958 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0b9b45f1-f282-4118-b627-94e4f8167d00-etc-machine-id\") pod \"manila-api-2\" (UID: \"0b9b45f1-f282-4118-b627-94e4f8167d00\") " pod="manila-kuttl-tests/manila-api-2" Feb 02 15:35:34 crc kubenswrapper[4733]: I0202 15:35:34.567951 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e2b1c4e-2735-437e-b2a1-482c2f909d8f-config-data\") pod \"manila-api-1\" (UID: \"2e2b1c4e-2735-437e-b2a1-482c2f909d8f\") " pod="manila-kuttl-tests/manila-api-1" Feb 02 15:35:34 crc kubenswrapper[4733]: I0202 15:35:34.568017 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b9b45f1-f282-4118-b627-94e4f8167d00-logs\") pod \"manila-api-2\" (UID: \"0b9b45f1-f282-4118-b627-94e4f8167d00\") " pod="manila-kuttl-tests/manila-api-2" Feb 02 15:35:34 crc kubenswrapper[4733]: I0202 15:35:34.568052 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e2b1c4e-2735-437e-b2a1-482c2f909d8f-scripts\") pod \"manila-api-1\" (UID: \"2e2b1c4e-2735-437e-b2a1-482c2f909d8f\") " pod="manila-kuttl-tests/manila-api-1" Feb 02 15:35:34 crc kubenswrapper[4733]: I0202 15:35:34.568079 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2e2b1c4e-2735-437e-b2a1-482c2f909d8f-config-data-custom\") pod \"manila-api-1\" (UID: \"2e2b1c4e-2735-437e-b2a1-482c2f909d8f\") " pod="manila-kuttl-tests/manila-api-1" Feb 02 15:35:34 crc kubenswrapper[4733]: I0202 15:35:34.568132 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2e2b1c4e-2735-437e-b2a1-482c2f909d8f-etc-machine-id\") pod \"manila-api-1\" (UID: \"2e2b1c4e-2735-437e-b2a1-482c2f909d8f\") " pod="manila-kuttl-tests/manila-api-1" Feb 02 15:35:34 crc kubenswrapper[4733]: I0202 15:35:34.568204 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkkfp\" (UniqueName: \"kubernetes.io/projected/0b9b45f1-f282-4118-b627-94e4f8167d00-kube-api-access-nkkfp\") pod \"manila-api-2\" (UID: \"0b9b45f1-f282-4118-b627-94e4f8167d00\") " pod="manila-kuttl-tests/manila-api-2" Feb 02 15:35:34 crc kubenswrapper[4733]: I0202 15:35:34.568268 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e2b1c4e-2735-437e-b2a1-482c2f909d8f-logs\") pod \"manila-api-1\" (UID: \"2e2b1c4e-2735-437e-b2a1-482c2f909d8f\") " pod="manila-kuttl-tests/manila-api-1" Feb 02 15:35:34 crc kubenswrapper[4733]: I0202 15:35:34.568309 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0b9b45f1-f282-4118-b627-94e4f8167d00-config-data-custom\") pod \"manila-api-2\" (UID: \"0b9b45f1-f282-4118-b627-94e4f8167d00\") " pod="manila-kuttl-tests/manila-api-2" Feb 02 15:35:34 crc kubenswrapper[4733]: I0202 15:35:34.568350 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b9b45f1-f282-4118-b627-94e4f8167d00-config-data\") pod \"manila-api-2\" (UID: \"0b9b45f1-f282-4118-b627-94e4f8167d00\") " pod="manila-kuttl-tests/manila-api-2" Feb 02 15:35:34 crc kubenswrapper[4733]: I0202 15:35:34.568389 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b9b45f1-f282-4118-b627-94e4f8167d00-scripts\") pod \"manila-api-2\" (UID: \"0b9b45f1-f282-4118-b627-94e4f8167d00\") " pod="manila-kuttl-tests/manila-api-2" Feb 02 15:35:34 crc kubenswrapper[4733]: I0202 15:35:34.568434 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5wkc\" (UniqueName: \"kubernetes.io/projected/2e2b1c4e-2735-437e-b2a1-482c2f909d8f-kube-api-access-x5wkc\") pod \"manila-api-1\" (UID: \"2e2b1c4e-2735-437e-b2a1-482c2f909d8f\") " pod="manila-kuttl-tests/manila-api-1" Feb 02 15:35:34 crc kubenswrapper[4733]: I0202 15:35:34.568480 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0b9b45f1-f282-4118-b627-94e4f8167d00-etc-machine-id\") pod \"manila-api-2\" (UID: \"0b9b45f1-f282-4118-b627-94e4f8167d00\") " pod="manila-kuttl-tests/manila-api-2" Feb 02 15:35:34 crc kubenswrapper[4733]: I0202 15:35:34.568595 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0b9b45f1-f282-4118-b627-94e4f8167d00-etc-machine-id\") pod \"manila-api-2\" (UID: \"0b9b45f1-f282-4118-b627-94e4f8167d00\") " pod="manila-kuttl-tests/manila-api-2" Feb 02 15:35:34 crc kubenswrapper[4733]: I0202 15:35:34.568638 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b9b45f1-f282-4118-b627-94e4f8167d00-logs\") pod \"manila-api-2\" (UID: \"0b9b45f1-f282-4118-b627-94e4f8167d00\") " pod="manila-kuttl-tests/manila-api-2" Feb 02 15:35:34 crc kubenswrapper[4733]: I0202 15:35:34.568928 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2e2b1c4e-2735-437e-b2a1-482c2f909d8f-etc-machine-id\") pod \"manila-api-1\" (UID: \"2e2b1c4e-2735-437e-b2a1-482c2f909d8f\") " pod="manila-kuttl-tests/manila-api-1" Feb 02 15:35:34 crc kubenswrapper[4733]: I0202 15:35:34.569078 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e2b1c4e-2735-437e-b2a1-482c2f909d8f-logs\") pod \"manila-api-1\" (UID: \"2e2b1c4e-2735-437e-b2a1-482c2f909d8f\") " pod="manila-kuttl-tests/manila-api-1" Feb 02 15:35:34 crc kubenswrapper[4733]: I0202 15:35:34.576775 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e2b1c4e-2735-437e-b2a1-482c2f909d8f-config-data\") pod \"manila-api-1\" (UID: \"2e2b1c4e-2735-437e-b2a1-482c2f909d8f\") " pod="manila-kuttl-tests/manila-api-1" Feb 02 15:35:34 crc kubenswrapper[4733]: I0202 15:35:34.576892 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b9b45f1-f282-4118-b627-94e4f8167d00-config-data\") pod \"manila-api-2\" (UID: \"0b9b45f1-f282-4118-b627-94e4f8167d00\") " pod="manila-kuttl-tests/manila-api-2" Feb 02 15:35:34 crc kubenswrapper[4733]: I0202 15:35:34.577942 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0b9b45f1-f282-4118-b627-94e4f8167d00-config-data-custom\") pod \"manila-api-2\" (UID: \"0b9b45f1-f282-4118-b627-94e4f8167d00\") " pod="manila-kuttl-tests/manila-api-2" Feb 02 15:35:34 crc kubenswrapper[4733]: I0202 15:35:34.580047 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2e2b1c4e-2735-437e-b2a1-482c2f909d8f-config-data-custom\") pod \"manila-api-1\" (UID: \"2e2b1c4e-2735-437e-b2a1-482c2f909d8f\") " pod="manila-kuttl-tests/manila-api-1" Feb 02 15:35:34 crc kubenswrapper[4733]: I0202 15:35:34.582450 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b9b45f1-f282-4118-b627-94e4f8167d00-scripts\") pod \"manila-api-2\" (UID: \"0b9b45f1-f282-4118-b627-94e4f8167d00\") " pod="manila-kuttl-tests/manila-api-2" Feb 02 15:35:34 crc kubenswrapper[4733]: I0202 15:35:34.587208 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5wkc\" (UniqueName: \"kubernetes.io/projected/2e2b1c4e-2735-437e-b2a1-482c2f909d8f-kube-api-access-x5wkc\") pod \"manila-api-1\" (UID: \"2e2b1c4e-2735-437e-b2a1-482c2f909d8f\") " pod="manila-kuttl-tests/manila-api-1" Feb 02 15:35:34 crc kubenswrapper[4733]: I0202 15:35:34.588478 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e2b1c4e-2735-437e-b2a1-482c2f909d8f-scripts\") pod \"manila-api-1\" (UID: \"2e2b1c4e-2735-437e-b2a1-482c2f909d8f\") " pod="manila-kuttl-tests/manila-api-1" Feb 02 15:35:34 crc kubenswrapper[4733]: I0202 15:35:34.595030 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkkfp\" (UniqueName: \"kubernetes.io/projected/0b9b45f1-f282-4118-b627-94e4f8167d00-kube-api-access-nkkfp\") pod \"manila-api-2\" (UID: \"0b9b45f1-f282-4118-b627-94e4f8167d00\") " pod="manila-kuttl-tests/manila-api-2" Feb 02 15:35:34 crc kubenswrapper[4733]: I0202 15:35:34.713949 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-api-2" Feb 02 15:35:34 crc kubenswrapper[4733]: I0202 15:35:34.730592 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-api-1" Feb 02 15:35:34 crc kubenswrapper[4733]: I0202 15:35:34.987767 4733 patch_prober.go:28] interesting pod/machine-config-daemon-zgsv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 15:35:34 crc kubenswrapper[4733]: I0202 15:35:34.987834 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 15:35:35 crc kubenswrapper[4733]: I0202 15:35:35.186247 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/manila-api-2"] Feb 02 15:35:35 crc kubenswrapper[4733]: I0202 15:35:35.234991 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/manila-api-1"] Feb 02 15:35:35 crc kubenswrapper[4733]: W0202 15:35:35.236308 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2e2b1c4e_2735_437e_b2a1_482c2f909d8f.slice/crio-ef5f49f6af61f5bb097f7b028c27e2a6dca412e11df1e50855644f4593edb609 WatchSource:0}: Error finding container ef5f49f6af61f5bb097f7b028c27e2a6dca412e11df1e50855644f4593edb609: Status 404 returned error can't find the container with id ef5f49f6af61f5bb097f7b028c27e2a6dca412e11df1e50855644f4593edb609 Feb 02 15:35:35 crc kubenswrapper[4733]: I0202 15:35:35.709518 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-api-1" event={"ID":"2e2b1c4e-2735-437e-b2a1-482c2f909d8f","Type":"ContainerStarted","Data":"6acafed3492f4b86621074c366a3a1ad688c73bab65f71c5ee885aa299cdd328"} Feb 02 15:35:35 crc kubenswrapper[4733]: I0202 15:35:35.710022 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-api-1" event={"ID":"2e2b1c4e-2735-437e-b2a1-482c2f909d8f","Type":"ContainerStarted","Data":"ef5f49f6af61f5bb097f7b028c27e2a6dca412e11df1e50855644f4593edb609"} Feb 02 15:35:35 crc kubenswrapper[4733]: I0202 15:35:35.712343 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-api-2" event={"ID":"0b9b45f1-f282-4118-b627-94e4f8167d00","Type":"ContainerStarted","Data":"5070b6e1f664efa353817064e7561412d758439b85ef3e6734829484575dc71e"} Feb 02 15:35:35 crc kubenswrapper[4733]: I0202 15:35:35.712381 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-api-2" event={"ID":"0b9b45f1-f282-4118-b627-94e4f8167d00","Type":"ContainerStarted","Data":"306877bd33623053c6d41b7d5342938505911cc441ecd30a41a3fce215293502"} Feb 02 15:35:36 crc kubenswrapper[4733]: I0202 15:35:36.722329 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-api-2" event={"ID":"0b9b45f1-f282-4118-b627-94e4f8167d00","Type":"ContainerStarted","Data":"19aafdf8df8c422fe422aa276d6b9d53ee049461160bdc86c6a58308ac6898fe"} Feb 02 15:35:36 crc kubenswrapper[4733]: I0202 15:35:36.722682 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="manila-kuttl-tests/manila-api-2" Feb 02 15:35:36 crc kubenswrapper[4733]: I0202 15:35:36.725237 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-api-1" event={"ID":"2e2b1c4e-2735-437e-b2a1-482c2f909d8f","Type":"ContainerStarted","Data":"3629e30f4e1d0eab923e142c1662f5bb743d3c3e0ad6d2cd38bdc3a9a7725367"} Feb 02 15:35:36 crc kubenswrapper[4733]: I0202 15:35:36.725842 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="manila-kuttl-tests/manila-api-1" Feb 02 15:35:36 crc kubenswrapper[4733]: I0202 15:35:36.743069 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="manila-kuttl-tests/manila-api-2" podStartSLOduration=2.7430522269999997 podStartE2EDuration="2.743052227s" podCreationTimestamp="2026-02-02 15:35:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:35:36.742302727 +0000 UTC m=+1220.193764085" watchObservedRunningTime="2026-02-02 15:35:36.743052227 +0000 UTC m=+1220.194513585" Feb 02 15:35:36 crc kubenswrapper[4733]: I0202 15:35:36.778138 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="manila-kuttl-tests/manila-api-1" podStartSLOduration=2.778105943 podStartE2EDuration="2.778105943s" podCreationTimestamp="2026-02-02 15:35:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:35:36.765853139 +0000 UTC m=+1220.217314567" watchObservedRunningTime="2026-02-02 15:35:36.778105943 +0000 UTC m=+1220.229567351" Feb 02 15:35:55 crc kubenswrapper[4733]: I0202 15:35:55.994897 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="manila-kuttl-tests/manila-api-1" Feb 02 15:35:56 crc kubenswrapper[4733]: I0202 15:35:56.007296 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="manila-kuttl-tests/manila-api-2" Feb 02 15:35:57 crc kubenswrapper[4733]: I0202 15:35:57.293811 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/manila-api-2"] Feb 02 15:35:57 crc kubenswrapper[4733]: I0202 15:35:57.294091 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="manila-kuttl-tests/manila-api-2" podUID="0b9b45f1-f282-4118-b627-94e4f8167d00" containerName="manila-api-log" containerID="cri-o://5070b6e1f664efa353817064e7561412d758439b85ef3e6734829484575dc71e" gracePeriod=30 Feb 02 15:35:57 crc kubenswrapper[4733]: I0202 15:35:57.294179 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="manila-kuttl-tests/manila-api-2" podUID="0b9b45f1-f282-4118-b627-94e4f8167d00" containerName="manila-api" containerID="cri-o://19aafdf8df8c422fe422aa276d6b9d53ee049461160bdc86c6a58308ac6898fe" gracePeriod=30 Feb 02 15:35:57 crc kubenswrapper[4733]: I0202 15:35:57.305237 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/manila-api-1"] Feb 02 15:35:57 crc kubenswrapper[4733]: I0202 15:35:57.305760 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="manila-kuttl-tests/manila-api-1" podUID="2e2b1c4e-2735-437e-b2a1-482c2f909d8f" containerName="manila-api-log" containerID="cri-o://6acafed3492f4b86621074c366a3a1ad688c73bab65f71c5ee885aa299cdd328" gracePeriod=30 Feb 02 15:35:57 crc kubenswrapper[4733]: I0202 15:35:57.305953 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="manila-kuttl-tests/manila-api-1" podUID="2e2b1c4e-2735-437e-b2a1-482c2f909d8f" containerName="manila-api" containerID="cri-o://3629e30f4e1d0eab923e142c1662f5bb743d3c3e0ad6d2cd38bdc3a9a7725367" gracePeriod=30 Feb 02 15:35:57 crc kubenswrapper[4733]: I0202 15:35:57.939456 4733 generic.go:334] "Generic (PLEG): container finished" podID="0b9b45f1-f282-4118-b627-94e4f8167d00" containerID="5070b6e1f664efa353817064e7561412d758439b85ef3e6734829484575dc71e" exitCode=143 Feb 02 15:35:57 crc kubenswrapper[4733]: I0202 15:35:57.939528 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-api-2" event={"ID":"0b9b45f1-f282-4118-b627-94e4f8167d00","Type":"ContainerDied","Data":"5070b6e1f664efa353817064e7561412d758439b85ef3e6734829484575dc71e"} Feb 02 15:35:57 crc kubenswrapper[4733]: I0202 15:35:57.942101 4733 generic.go:334] "Generic (PLEG): container finished" podID="2e2b1c4e-2735-437e-b2a1-482c2f909d8f" containerID="6acafed3492f4b86621074c366a3a1ad688c73bab65f71c5ee885aa299cdd328" exitCode=143 Feb 02 15:35:57 crc kubenswrapper[4733]: I0202 15:35:57.942204 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-api-1" event={"ID":"2e2b1c4e-2735-437e-b2a1-482c2f909d8f","Type":"ContainerDied","Data":"6acafed3492f4b86621074c366a3a1ad688c73bab65f71c5ee885aa299cdd328"} Feb 02 15:36:00 crc kubenswrapper[4733]: I0202 15:36:00.810514 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-api-2" Feb 02 15:36:00 crc kubenswrapper[4733]: I0202 15:36:00.900195 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-api-1" Feb 02 15:36:00 crc kubenswrapper[4733]: I0202 15:36:00.903672 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b9b45f1-f282-4118-b627-94e4f8167d00-logs\") pod \"0b9b45f1-f282-4118-b627-94e4f8167d00\" (UID: \"0b9b45f1-f282-4118-b627-94e4f8167d00\") " Feb 02 15:36:00 crc kubenswrapper[4733]: I0202 15:36:00.903780 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b9b45f1-f282-4118-b627-94e4f8167d00-scripts\") pod \"0b9b45f1-f282-4118-b627-94e4f8167d00\" (UID: \"0b9b45f1-f282-4118-b627-94e4f8167d00\") " Feb 02 15:36:00 crc kubenswrapper[4733]: I0202 15:36:00.903837 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nkkfp\" (UniqueName: \"kubernetes.io/projected/0b9b45f1-f282-4118-b627-94e4f8167d00-kube-api-access-nkkfp\") pod \"0b9b45f1-f282-4118-b627-94e4f8167d00\" (UID: \"0b9b45f1-f282-4118-b627-94e4f8167d00\") " Feb 02 15:36:00 crc kubenswrapper[4733]: I0202 15:36:00.903882 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0b9b45f1-f282-4118-b627-94e4f8167d00-config-data-custom\") pod \"0b9b45f1-f282-4118-b627-94e4f8167d00\" (UID: \"0b9b45f1-f282-4118-b627-94e4f8167d00\") " Feb 02 15:36:00 crc kubenswrapper[4733]: I0202 15:36:00.903912 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0b9b45f1-f282-4118-b627-94e4f8167d00-etc-machine-id\") pod \"0b9b45f1-f282-4118-b627-94e4f8167d00\" (UID: \"0b9b45f1-f282-4118-b627-94e4f8167d00\") " Feb 02 15:36:00 crc kubenswrapper[4733]: I0202 15:36:00.903943 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b9b45f1-f282-4118-b627-94e4f8167d00-config-data\") pod \"0b9b45f1-f282-4118-b627-94e4f8167d00\" (UID: \"0b9b45f1-f282-4118-b627-94e4f8167d00\") " Feb 02 15:36:00 crc kubenswrapper[4733]: I0202 15:36:00.904854 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b9b45f1-f282-4118-b627-94e4f8167d00-logs" (OuterVolumeSpecName: "logs") pod "0b9b45f1-f282-4118-b627-94e4f8167d00" (UID: "0b9b45f1-f282-4118-b627-94e4f8167d00"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:36:00 crc kubenswrapper[4733]: I0202 15:36:00.904907 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0b9b45f1-f282-4118-b627-94e4f8167d00-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "0b9b45f1-f282-4118-b627-94e4f8167d00" (UID: "0b9b45f1-f282-4118-b627-94e4f8167d00"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 15:36:00 crc kubenswrapper[4733]: I0202 15:36:00.911071 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b9b45f1-f282-4118-b627-94e4f8167d00-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "0b9b45f1-f282-4118-b627-94e4f8167d00" (UID: "0b9b45f1-f282-4118-b627-94e4f8167d00"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:36:00 crc kubenswrapper[4733]: I0202 15:36:00.911421 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b9b45f1-f282-4118-b627-94e4f8167d00-scripts" (OuterVolumeSpecName: "scripts") pod "0b9b45f1-f282-4118-b627-94e4f8167d00" (UID: "0b9b45f1-f282-4118-b627-94e4f8167d00"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:36:00 crc kubenswrapper[4733]: I0202 15:36:00.921016 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b9b45f1-f282-4118-b627-94e4f8167d00-kube-api-access-nkkfp" (OuterVolumeSpecName: "kube-api-access-nkkfp") pod "0b9b45f1-f282-4118-b627-94e4f8167d00" (UID: "0b9b45f1-f282-4118-b627-94e4f8167d00"). InnerVolumeSpecName "kube-api-access-nkkfp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:36:00 crc kubenswrapper[4733]: I0202 15:36:00.965906 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b9b45f1-f282-4118-b627-94e4f8167d00-config-data" (OuterVolumeSpecName: "config-data") pod "0b9b45f1-f282-4118-b627-94e4f8167d00" (UID: "0b9b45f1-f282-4118-b627-94e4f8167d00"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:36:00 crc kubenswrapper[4733]: I0202 15:36:00.968320 4733 generic.go:334] "Generic (PLEG): container finished" podID="0b9b45f1-f282-4118-b627-94e4f8167d00" containerID="19aafdf8df8c422fe422aa276d6b9d53ee049461160bdc86c6a58308ac6898fe" exitCode=0 Feb 02 15:36:00 crc kubenswrapper[4733]: I0202 15:36:00.968401 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-api-2" event={"ID":"0b9b45f1-f282-4118-b627-94e4f8167d00","Type":"ContainerDied","Data":"19aafdf8df8c422fe422aa276d6b9d53ee049461160bdc86c6a58308ac6898fe"} Feb 02 15:36:00 crc kubenswrapper[4733]: I0202 15:36:00.968418 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-api-2" Feb 02 15:36:00 crc kubenswrapper[4733]: I0202 15:36:00.968433 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-api-2" event={"ID":"0b9b45f1-f282-4118-b627-94e4f8167d00","Type":"ContainerDied","Data":"306877bd33623053c6d41b7d5342938505911cc441ecd30a41a3fce215293502"} Feb 02 15:36:00 crc kubenswrapper[4733]: I0202 15:36:00.968449 4733 scope.go:117] "RemoveContainer" containerID="19aafdf8df8c422fe422aa276d6b9d53ee049461160bdc86c6a58308ac6898fe" Feb 02 15:36:00 crc kubenswrapper[4733]: I0202 15:36:00.973251 4733 generic.go:334] "Generic (PLEG): container finished" podID="2e2b1c4e-2735-437e-b2a1-482c2f909d8f" containerID="3629e30f4e1d0eab923e142c1662f5bb743d3c3e0ad6d2cd38bdc3a9a7725367" exitCode=0 Feb 02 15:36:00 crc kubenswrapper[4733]: I0202 15:36:00.973291 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-api-1" event={"ID":"2e2b1c4e-2735-437e-b2a1-482c2f909d8f","Type":"ContainerDied","Data":"3629e30f4e1d0eab923e142c1662f5bb743d3c3e0ad6d2cd38bdc3a9a7725367"} Feb 02 15:36:00 crc kubenswrapper[4733]: I0202 15:36:00.973318 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-api-1" event={"ID":"2e2b1c4e-2735-437e-b2a1-482c2f909d8f","Type":"ContainerDied","Data":"ef5f49f6af61f5bb097f7b028c27e2a6dca412e11df1e50855644f4593edb609"} Feb 02 15:36:00 crc kubenswrapper[4733]: I0202 15:36:00.973370 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-api-1" Feb 02 15:36:00 crc kubenswrapper[4733]: I0202 15:36:00.988648 4733 scope.go:117] "RemoveContainer" containerID="5070b6e1f664efa353817064e7561412d758439b85ef3e6734829484575dc71e" Feb 02 15:36:01 crc kubenswrapper[4733]: I0202 15:36:01.004909 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/manila-api-2"] Feb 02 15:36:01 crc kubenswrapper[4733]: I0202 15:36:01.006402 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e2b1c4e-2735-437e-b2a1-482c2f909d8f-scripts\") pod \"2e2b1c4e-2735-437e-b2a1-482c2f909d8f\" (UID: \"2e2b1c4e-2735-437e-b2a1-482c2f909d8f\") " Feb 02 15:36:01 crc kubenswrapper[4733]: I0202 15:36:01.006443 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e2b1c4e-2735-437e-b2a1-482c2f909d8f-logs\") pod \"2e2b1c4e-2735-437e-b2a1-482c2f909d8f\" (UID: \"2e2b1c4e-2735-437e-b2a1-482c2f909d8f\") " Feb 02 15:36:01 crc kubenswrapper[4733]: I0202 15:36:01.006472 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2e2b1c4e-2735-437e-b2a1-482c2f909d8f-etc-machine-id\") pod \"2e2b1c4e-2735-437e-b2a1-482c2f909d8f\" (UID: \"2e2b1c4e-2735-437e-b2a1-482c2f909d8f\") " Feb 02 15:36:01 crc kubenswrapper[4733]: I0202 15:36:01.006517 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2e2b1c4e-2735-437e-b2a1-482c2f909d8f-config-data-custom\") pod \"2e2b1c4e-2735-437e-b2a1-482c2f909d8f\" (UID: \"2e2b1c4e-2735-437e-b2a1-482c2f909d8f\") " Feb 02 15:36:01 crc kubenswrapper[4733]: I0202 15:36:01.006565 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e2b1c4e-2735-437e-b2a1-482c2f909d8f-config-data\") pod \"2e2b1c4e-2735-437e-b2a1-482c2f909d8f\" (UID: \"2e2b1c4e-2735-437e-b2a1-482c2f909d8f\") " Feb 02 15:36:01 crc kubenswrapper[4733]: I0202 15:36:01.006589 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5wkc\" (UniqueName: \"kubernetes.io/projected/2e2b1c4e-2735-437e-b2a1-482c2f909d8f-kube-api-access-x5wkc\") pod \"2e2b1c4e-2735-437e-b2a1-482c2f909d8f\" (UID: \"2e2b1c4e-2735-437e-b2a1-482c2f909d8f\") " Feb 02 15:36:01 crc kubenswrapper[4733]: I0202 15:36:01.006896 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b9b45f1-f282-4118-b627-94e4f8167d00-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 15:36:01 crc kubenswrapper[4733]: I0202 15:36:01.006923 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nkkfp\" (UniqueName: \"kubernetes.io/projected/0b9b45f1-f282-4118-b627-94e4f8167d00-kube-api-access-nkkfp\") on node \"crc\" DevicePath \"\"" Feb 02 15:36:01 crc kubenswrapper[4733]: I0202 15:36:01.006942 4733 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0b9b45f1-f282-4118-b627-94e4f8167d00-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 02 15:36:01 crc kubenswrapper[4733]: I0202 15:36:01.006955 4733 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0b9b45f1-f282-4118-b627-94e4f8167d00-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 02 15:36:01 crc kubenswrapper[4733]: I0202 15:36:01.006966 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b9b45f1-f282-4118-b627-94e4f8167d00-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 15:36:01 crc kubenswrapper[4733]: I0202 15:36:01.006977 4733 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b9b45f1-f282-4118-b627-94e4f8167d00-logs\") on node \"crc\" DevicePath \"\"" Feb 02 15:36:01 crc kubenswrapper[4733]: I0202 15:36:01.007085 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e2b1c4e-2735-437e-b2a1-482c2f909d8f-logs" (OuterVolumeSpecName: "logs") pod "2e2b1c4e-2735-437e-b2a1-482c2f909d8f" (UID: "2e2b1c4e-2735-437e-b2a1-482c2f909d8f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:36:01 crc kubenswrapper[4733]: I0202 15:36:01.007430 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2e2b1c4e-2735-437e-b2a1-482c2f909d8f-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "2e2b1c4e-2735-437e-b2a1-482c2f909d8f" (UID: "2e2b1c4e-2735-437e-b2a1-482c2f909d8f"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 15:36:01 crc kubenswrapper[4733]: I0202 15:36:01.009477 4733 scope.go:117] "RemoveContainer" containerID="19aafdf8df8c422fe422aa276d6b9d53ee049461160bdc86c6a58308ac6898fe" Feb 02 15:36:01 crc kubenswrapper[4733]: I0202 15:36:01.009782 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e2b1c4e-2735-437e-b2a1-482c2f909d8f-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "2e2b1c4e-2735-437e-b2a1-482c2f909d8f" (UID: "2e2b1c4e-2735-437e-b2a1-482c2f909d8f"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:36:01 crc kubenswrapper[4733]: E0202 15:36:01.009849 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19aafdf8df8c422fe422aa276d6b9d53ee049461160bdc86c6a58308ac6898fe\": container with ID starting with 19aafdf8df8c422fe422aa276d6b9d53ee049461160bdc86c6a58308ac6898fe not found: ID does not exist" containerID="19aafdf8df8c422fe422aa276d6b9d53ee049461160bdc86c6a58308ac6898fe" Feb 02 15:36:01 crc kubenswrapper[4733]: I0202 15:36:01.009849 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e2b1c4e-2735-437e-b2a1-482c2f909d8f-scripts" (OuterVolumeSpecName: "scripts") pod "2e2b1c4e-2735-437e-b2a1-482c2f909d8f" (UID: "2e2b1c4e-2735-437e-b2a1-482c2f909d8f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:36:01 crc kubenswrapper[4733]: I0202 15:36:01.009884 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19aafdf8df8c422fe422aa276d6b9d53ee049461160bdc86c6a58308ac6898fe"} err="failed to get container status \"19aafdf8df8c422fe422aa276d6b9d53ee049461160bdc86c6a58308ac6898fe\": rpc error: code = NotFound desc = could not find container \"19aafdf8df8c422fe422aa276d6b9d53ee049461160bdc86c6a58308ac6898fe\": container with ID starting with 19aafdf8df8c422fe422aa276d6b9d53ee049461160bdc86c6a58308ac6898fe not found: ID does not exist" Feb 02 15:36:01 crc kubenswrapper[4733]: I0202 15:36:01.009914 4733 scope.go:117] "RemoveContainer" containerID="5070b6e1f664efa353817064e7561412d758439b85ef3e6734829484575dc71e" Feb 02 15:36:01 crc kubenswrapper[4733]: E0202 15:36:01.010446 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5070b6e1f664efa353817064e7561412d758439b85ef3e6734829484575dc71e\": container with ID starting with 5070b6e1f664efa353817064e7561412d758439b85ef3e6734829484575dc71e not found: ID does not exist" containerID="5070b6e1f664efa353817064e7561412d758439b85ef3e6734829484575dc71e" Feb 02 15:36:01 crc kubenswrapper[4733]: I0202 15:36:01.010495 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5070b6e1f664efa353817064e7561412d758439b85ef3e6734829484575dc71e"} err="failed to get container status \"5070b6e1f664efa353817064e7561412d758439b85ef3e6734829484575dc71e\": rpc error: code = NotFound desc = could not find container \"5070b6e1f664efa353817064e7561412d758439b85ef3e6734829484575dc71e\": container with ID starting with 5070b6e1f664efa353817064e7561412d758439b85ef3e6734829484575dc71e not found: ID does not exist" Feb 02 15:36:01 crc kubenswrapper[4733]: I0202 15:36:01.010609 4733 scope.go:117] "RemoveContainer" containerID="3629e30f4e1d0eab923e142c1662f5bb743d3c3e0ad6d2cd38bdc3a9a7725367" Feb 02 15:36:01 crc kubenswrapper[4733]: I0202 15:36:01.010888 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["manila-kuttl-tests/manila-api-2"] Feb 02 15:36:01 crc kubenswrapper[4733]: I0202 15:36:01.011080 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e2b1c4e-2735-437e-b2a1-482c2f909d8f-kube-api-access-x5wkc" (OuterVolumeSpecName: "kube-api-access-x5wkc") pod "2e2b1c4e-2735-437e-b2a1-482c2f909d8f" (UID: "2e2b1c4e-2735-437e-b2a1-482c2f909d8f"). InnerVolumeSpecName "kube-api-access-x5wkc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:36:01 crc kubenswrapper[4733]: I0202 15:36:01.027235 4733 scope.go:117] "RemoveContainer" containerID="6acafed3492f4b86621074c366a3a1ad688c73bab65f71c5ee885aa299cdd328" Feb 02 15:36:01 crc kubenswrapper[4733]: I0202 15:36:01.042061 4733 scope.go:117] "RemoveContainer" containerID="3629e30f4e1d0eab923e142c1662f5bb743d3c3e0ad6d2cd38bdc3a9a7725367" Feb 02 15:36:01 crc kubenswrapper[4733]: E0202 15:36:01.042491 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3629e30f4e1d0eab923e142c1662f5bb743d3c3e0ad6d2cd38bdc3a9a7725367\": container with ID starting with 3629e30f4e1d0eab923e142c1662f5bb743d3c3e0ad6d2cd38bdc3a9a7725367 not found: ID does not exist" containerID="3629e30f4e1d0eab923e142c1662f5bb743d3c3e0ad6d2cd38bdc3a9a7725367" Feb 02 15:36:01 crc kubenswrapper[4733]: I0202 15:36:01.042541 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3629e30f4e1d0eab923e142c1662f5bb743d3c3e0ad6d2cd38bdc3a9a7725367"} err="failed to get container status \"3629e30f4e1d0eab923e142c1662f5bb743d3c3e0ad6d2cd38bdc3a9a7725367\": rpc error: code = NotFound desc = could not find container \"3629e30f4e1d0eab923e142c1662f5bb743d3c3e0ad6d2cd38bdc3a9a7725367\": container with ID starting with 3629e30f4e1d0eab923e142c1662f5bb743d3c3e0ad6d2cd38bdc3a9a7725367 not found: ID does not exist" Feb 02 15:36:01 crc kubenswrapper[4733]: I0202 15:36:01.042574 4733 scope.go:117] "RemoveContainer" containerID="6acafed3492f4b86621074c366a3a1ad688c73bab65f71c5ee885aa299cdd328" Feb 02 15:36:01 crc kubenswrapper[4733]: E0202 15:36:01.043053 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6acafed3492f4b86621074c366a3a1ad688c73bab65f71c5ee885aa299cdd328\": container with ID starting with 6acafed3492f4b86621074c366a3a1ad688c73bab65f71c5ee885aa299cdd328 not found: ID does not exist" containerID="6acafed3492f4b86621074c366a3a1ad688c73bab65f71c5ee885aa299cdd328" Feb 02 15:36:01 crc kubenswrapper[4733]: I0202 15:36:01.043103 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6acafed3492f4b86621074c366a3a1ad688c73bab65f71c5ee885aa299cdd328"} err="failed to get container status \"6acafed3492f4b86621074c366a3a1ad688c73bab65f71c5ee885aa299cdd328\": rpc error: code = NotFound desc = could not find container \"6acafed3492f4b86621074c366a3a1ad688c73bab65f71c5ee885aa299cdd328\": container with ID starting with 6acafed3492f4b86621074c366a3a1ad688c73bab65f71c5ee885aa299cdd328 not found: ID does not exist" Feb 02 15:36:01 crc kubenswrapper[4733]: I0202 15:36:01.048590 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e2b1c4e-2735-437e-b2a1-482c2f909d8f-config-data" (OuterVolumeSpecName: "config-data") pod "2e2b1c4e-2735-437e-b2a1-482c2f909d8f" (UID: "2e2b1c4e-2735-437e-b2a1-482c2f909d8f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:36:01 crc kubenswrapper[4733]: I0202 15:36:01.108102 4733 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2e2b1c4e-2735-437e-b2a1-482c2f909d8f-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 02 15:36:01 crc kubenswrapper[4733]: I0202 15:36:01.108143 4733 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2e2b1c4e-2735-437e-b2a1-482c2f909d8f-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 02 15:36:01 crc kubenswrapper[4733]: I0202 15:36:01.108159 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e2b1c4e-2735-437e-b2a1-482c2f909d8f-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 15:36:01 crc kubenswrapper[4733]: I0202 15:36:01.108174 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5wkc\" (UniqueName: \"kubernetes.io/projected/2e2b1c4e-2735-437e-b2a1-482c2f909d8f-kube-api-access-x5wkc\") on node \"crc\" DevicePath \"\"" Feb 02 15:36:01 crc kubenswrapper[4733]: I0202 15:36:01.108201 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e2b1c4e-2735-437e-b2a1-482c2f909d8f-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 15:36:01 crc kubenswrapper[4733]: I0202 15:36:01.108214 4733 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e2b1c4e-2735-437e-b2a1-482c2f909d8f-logs\") on node \"crc\" DevicePath \"\"" Feb 02 15:36:01 crc kubenswrapper[4733]: I0202 15:36:01.267366 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b9b45f1-f282-4118-b627-94e4f8167d00" path="/var/lib/kubelet/pods/0b9b45f1-f282-4118-b627-94e4f8167d00/volumes" Feb 02 15:36:01 crc kubenswrapper[4733]: I0202 15:36:01.314605 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/manila-api-1"] Feb 02 15:36:01 crc kubenswrapper[4733]: I0202 15:36:01.322757 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["manila-kuttl-tests/manila-api-1"] Feb 02 15:36:02 crc kubenswrapper[4733]: I0202 15:36:02.740034 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["manila-kuttl-tests/manila-scheduler-1"] Feb 02 15:36:02 crc kubenswrapper[4733]: E0202 15:36:02.740543 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e2b1c4e-2735-437e-b2a1-482c2f909d8f" containerName="manila-api" Feb 02 15:36:02 crc kubenswrapper[4733]: I0202 15:36:02.740556 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e2b1c4e-2735-437e-b2a1-482c2f909d8f" containerName="manila-api" Feb 02 15:36:02 crc kubenswrapper[4733]: E0202 15:36:02.740565 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e2b1c4e-2735-437e-b2a1-482c2f909d8f" containerName="manila-api-log" Feb 02 15:36:02 crc kubenswrapper[4733]: I0202 15:36:02.740571 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e2b1c4e-2735-437e-b2a1-482c2f909d8f" containerName="manila-api-log" Feb 02 15:36:02 crc kubenswrapper[4733]: E0202 15:36:02.740583 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b9b45f1-f282-4118-b627-94e4f8167d00" containerName="manila-api-log" Feb 02 15:36:02 crc kubenswrapper[4733]: I0202 15:36:02.740588 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b9b45f1-f282-4118-b627-94e4f8167d00" containerName="manila-api-log" Feb 02 15:36:02 crc kubenswrapper[4733]: E0202 15:36:02.740597 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b9b45f1-f282-4118-b627-94e4f8167d00" containerName="manila-api" Feb 02 15:36:02 crc kubenswrapper[4733]: I0202 15:36:02.740603 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b9b45f1-f282-4118-b627-94e4f8167d00" containerName="manila-api" Feb 02 15:36:02 crc kubenswrapper[4733]: I0202 15:36:02.740733 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b9b45f1-f282-4118-b627-94e4f8167d00" containerName="manila-api" Feb 02 15:36:02 crc kubenswrapper[4733]: I0202 15:36:02.740747 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b9b45f1-f282-4118-b627-94e4f8167d00" containerName="manila-api-log" Feb 02 15:36:02 crc kubenswrapper[4733]: I0202 15:36:02.740755 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e2b1c4e-2735-437e-b2a1-482c2f909d8f" containerName="manila-api" Feb 02 15:36:02 crc kubenswrapper[4733]: I0202 15:36:02.740764 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e2b1c4e-2735-437e-b2a1-482c2f909d8f" containerName="manila-api-log" Feb 02 15:36:02 crc kubenswrapper[4733]: I0202 15:36:02.741478 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-scheduler-1" Feb 02 15:36:02 crc kubenswrapper[4733]: I0202 15:36:02.750278 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/manila-scheduler-1"] Feb 02 15:36:02 crc kubenswrapper[4733]: I0202 15:36:02.832392 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78bb94e9-3aad-4917-8a42-105b948191a6-scripts\") pod \"manila-scheduler-1\" (UID: \"78bb94e9-3aad-4917-8a42-105b948191a6\") " pod="manila-kuttl-tests/manila-scheduler-1" Feb 02 15:36:02 crc kubenswrapper[4733]: I0202 15:36:02.832456 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkv4z\" (UniqueName: \"kubernetes.io/projected/78bb94e9-3aad-4917-8a42-105b948191a6-kube-api-access-fkv4z\") pod \"manila-scheduler-1\" (UID: \"78bb94e9-3aad-4917-8a42-105b948191a6\") " pod="manila-kuttl-tests/manila-scheduler-1" Feb 02 15:36:02 crc kubenswrapper[4733]: I0202 15:36:02.832480 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78bb94e9-3aad-4917-8a42-105b948191a6-config-data\") pod \"manila-scheduler-1\" (UID: \"78bb94e9-3aad-4917-8a42-105b948191a6\") " pod="manila-kuttl-tests/manila-scheduler-1" Feb 02 15:36:02 crc kubenswrapper[4733]: I0202 15:36:02.832555 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/78bb94e9-3aad-4917-8a42-105b948191a6-etc-machine-id\") pod \"manila-scheduler-1\" (UID: \"78bb94e9-3aad-4917-8a42-105b948191a6\") " pod="manila-kuttl-tests/manila-scheduler-1" Feb 02 15:36:02 crc kubenswrapper[4733]: I0202 15:36:02.832578 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/78bb94e9-3aad-4917-8a42-105b948191a6-config-data-custom\") pod \"manila-scheduler-1\" (UID: \"78bb94e9-3aad-4917-8a42-105b948191a6\") " pod="manila-kuttl-tests/manila-scheduler-1" Feb 02 15:36:02 crc kubenswrapper[4733]: I0202 15:36:02.934093 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/78bb94e9-3aad-4917-8a42-105b948191a6-etc-machine-id\") pod \"manila-scheduler-1\" (UID: \"78bb94e9-3aad-4917-8a42-105b948191a6\") " pod="manila-kuttl-tests/manila-scheduler-1" Feb 02 15:36:02 crc kubenswrapper[4733]: I0202 15:36:02.934192 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/78bb94e9-3aad-4917-8a42-105b948191a6-config-data-custom\") pod \"manila-scheduler-1\" (UID: \"78bb94e9-3aad-4917-8a42-105b948191a6\") " pod="manila-kuttl-tests/manila-scheduler-1" Feb 02 15:36:02 crc kubenswrapper[4733]: I0202 15:36:02.934231 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78bb94e9-3aad-4917-8a42-105b948191a6-scripts\") pod \"manila-scheduler-1\" (UID: \"78bb94e9-3aad-4917-8a42-105b948191a6\") " pod="manila-kuttl-tests/manila-scheduler-1" Feb 02 15:36:02 crc kubenswrapper[4733]: I0202 15:36:02.934249 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/78bb94e9-3aad-4917-8a42-105b948191a6-etc-machine-id\") pod \"manila-scheduler-1\" (UID: \"78bb94e9-3aad-4917-8a42-105b948191a6\") " pod="manila-kuttl-tests/manila-scheduler-1" Feb 02 15:36:02 crc kubenswrapper[4733]: I0202 15:36:02.934286 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkv4z\" (UniqueName: \"kubernetes.io/projected/78bb94e9-3aad-4917-8a42-105b948191a6-kube-api-access-fkv4z\") pod \"manila-scheduler-1\" (UID: \"78bb94e9-3aad-4917-8a42-105b948191a6\") " pod="manila-kuttl-tests/manila-scheduler-1" Feb 02 15:36:02 crc kubenswrapper[4733]: I0202 15:36:02.934326 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78bb94e9-3aad-4917-8a42-105b948191a6-config-data\") pod \"manila-scheduler-1\" (UID: \"78bb94e9-3aad-4917-8a42-105b948191a6\") " pod="manila-kuttl-tests/manila-scheduler-1" Feb 02 15:36:02 crc kubenswrapper[4733]: I0202 15:36:02.940124 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/78bb94e9-3aad-4917-8a42-105b948191a6-config-data-custom\") pod \"manila-scheduler-1\" (UID: \"78bb94e9-3aad-4917-8a42-105b948191a6\") " pod="manila-kuttl-tests/manila-scheduler-1" Feb 02 15:36:02 crc kubenswrapper[4733]: I0202 15:36:02.945783 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78bb94e9-3aad-4917-8a42-105b948191a6-scripts\") pod \"manila-scheduler-1\" (UID: \"78bb94e9-3aad-4917-8a42-105b948191a6\") " pod="manila-kuttl-tests/manila-scheduler-1" Feb 02 15:36:02 crc kubenswrapper[4733]: I0202 15:36:02.946969 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78bb94e9-3aad-4917-8a42-105b948191a6-config-data\") pod \"manila-scheduler-1\" (UID: \"78bb94e9-3aad-4917-8a42-105b948191a6\") " pod="manila-kuttl-tests/manila-scheduler-1" Feb 02 15:36:02 crc kubenswrapper[4733]: I0202 15:36:02.949989 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkv4z\" (UniqueName: \"kubernetes.io/projected/78bb94e9-3aad-4917-8a42-105b948191a6-kube-api-access-fkv4z\") pod \"manila-scheduler-1\" (UID: \"78bb94e9-3aad-4917-8a42-105b948191a6\") " pod="manila-kuttl-tests/manila-scheduler-1" Feb 02 15:36:03 crc kubenswrapper[4733]: I0202 15:36:03.097211 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-scheduler-1" Feb 02 15:36:03 crc kubenswrapper[4733]: I0202 15:36:03.264125 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e2b1c4e-2735-437e-b2a1-482c2f909d8f" path="/var/lib/kubelet/pods/2e2b1c4e-2735-437e-b2a1-482c2f909d8f/volumes" Feb 02 15:36:03 crc kubenswrapper[4733]: I0202 15:36:03.937492 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/manila-scheduler-1"] Feb 02 15:36:03 crc kubenswrapper[4733]: W0202 15:36:03.949340 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod78bb94e9_3aad_4917_8a42_105b948191a6.slice/crio-d0efc3f8c2c8ee1132fc513c873eca127db5a1816c15b4348a02bcd87af57b76 WatchSource:0}: Error finding container d0efc3f8c2c8ee1132fc513c873eca127db5a1816c15b4348a02bcd87af57b76: Status 404 returned error can't find the container with id d0efc3f8c2c8ee1132fc513c873eca127db5a1816c15b4348a02bcd87af57b76 Feb 02 15:36:03 crc kubenswrapper[4733]: I0202 15:36:03.996697 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-scheduler-1" event={"ID":"78bb94e9-3aad-4917-8a42-105b948191a6","Type":"ContainerStarted","Data":"d0efc3f8c2c8ee1132fc513c873eca127db5a1816c15b4348a02bcd87af57b76"} Feb 02 15:36:04 crc kubenswrapper[4733]: I0202 15:36:04.988029 4733 patch_prober.go:28] interesting pod/machine-config-daemon-zgsv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 15:36:04 crc kubenswrapper[4733]: I0202 15:36:04.988667 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 15:36:05 crc kubenswrapper[4733]: I0202 15:36:05.005224 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-scheduler-1" event={"ID":"78bb94e9-3aad-4917-8a42-105b948191a6","Type":"ContainerStarted","Data":"c7f431f57e22c98695fc1863dff3e6d6c1ca0c21042e82a1b590b47d616d2bd7"} Feb 02 15:36:05 crc kubenswrapper[4733]: I0202 15:36:05.005269 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-scheduler-1" event={"ID":"78bb94e9-3aad-4917-8a42-105b948191a6","Type":"ContainerStarted","Data":"2ac63b9e1e48ebf555f89cf6b525c9b5d7b845fff893abba074691caf5188aaa"} Feb 02 15:36:05 crc kubenswrapper[4733]: I0202 15:36:05.024650 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="manila-kuttl-tests/manila-scheduler-1" podStartSLOduration=3.024155947 podStartE2EDuration="3.024155947s" podCreationTimestamp="2026-02-02 15:36:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:36:05.020672375 +0000 UTC m=+1248.472133733" watchObservedRunningTime="2026-02-02 15:36:05.024155947 +0000 UTC m=+1248.475617315" Feb 02 15:36:13 crc kubenswrapper[4733]: I0202 15:36:13.098235 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="manila-kuttl-tests/manila-scheduler-1" Feb 02 15:36:24 crc kubenswrapper[4733]: I0202 15:36:24.507793 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="manila-kuttl-tests/manila-scheduler-1" Feb 02 15:36:24 crc kubenswrapper[4733]: I0202 15:36:24.600915 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["manila-kuttl-tests/manila-scheduler-2"] Feb 02 15:36:24 crc kubenswrapper[4733]: I0202 15:36:24.602496 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-scheduler-2" Feb 02 15:36:24 crc kubenswrapper[4733]: I0202 15:36:24.625949 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/manila-scheduler-2"] Feb 02 15:36:24 crc kubenswrapper[4733]: I0202 15:36:24.699017 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7d57375e-7e2f-4870-8c60-555921f94b5e-config-data-custom\") pod \"manila-scheduler-2\" (UID: \"7d57375e-7e2f-4870-8c60-555921f94b5e\") " pod="manila-kuttl-tests/manila-scheduler-2" Feb 02 15:36:24 crc kubenswrapper[4733]: I0202 15:36:24.699072 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7d57375e-7e2f-4870-8c60-555921f94b5e-etc-machine-id\") pod \"manila-scheduler-2\" (UID: \"7d57375e-7e2f-4870-8c60-555921f94b5e\") " pod="manila-kuttl-tests/manila-scheduler-2" Feb 02 15:36:24 crc kubenswrapper[4733]: I0202 15:36:24.699094 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnn56\" (UniqueName: \"kubernetes.io/projected/7d57375e-7e2f-4870-8c60-555921f94b5e-kube-api-access-fnn56\") pod \"manila-scheduler-2\" (UID: \"7d57375e-7e2f-4870-8c60-555921f94b5e\") " pod="manila-kuttl-tests/manila-scheduler-2" Feb 02 15:36:24 crc kubenswrapper[4733]: I0202 15:36:24.699126 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d57375e-7e2f-4870-8c60-555921f94b5e-scripts\") pod \"manila-scheduler-2\" (UID: \"7d57375e-7e2f-4870-8c60-555921f94b5e\") " pod="manila-kuttl-tests/manila-scheduler-2" Feb 02 15:36:24 crc kubenswrapper[4733]: I0202 15:36:24.699179 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d57375e-7e2f-4870-8c60-555921f94b5e-config-data\") pod \"manila-scheduler-2\" (UID: \"7d57375e-7e2f-4870-8c60-555921f94b5e\") " pod="manila-kuttl-tests/manila-scheduler-2" Feb 02 15:36:24 crc kubenswrapper[4733]: I0202 15:36:24.800851 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7d57375e-7e2f-4870-8c60-555921f94b5e-config-data-custom\") pod \"manila-scheduler-2\" (UID: \"7d57375e-7e2f-4870-8c60-555921f94b5e\") " pod="manila-kuttl-tests/manila-scheduler-2" Feb 02 15:36:24 crc kubenswrapper[4733]: I0202 15:36:24.800943 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7d57375e-7e2f-4870-8c60-555921f94b5e-etc-machine-id\") pod \"manila-scheduler-2\" (UID: \"7d57375e-7e2f-4870-8c60-555921f94b5e\") " pod="manila-kuttl-tests/manila-scheduler-2" Feb 02 15:36:24 crc kubenswrapper[4733]: I0202 15:36:24.800986 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnn56\" (UniqueName: \"kubernetes.io/projected/7d57375e-7e2f-4870-8c60-555921f94b5e-kube-api-access-fnn56\") pod \"manila-scheduler-2\" (UID: \"7d57375e-7e2f-4870-8c60-555921f94b5e\") " pod="manila-kuttl-tests/manila-scheduler-2" Feb 02 15:36:24 crc kubenswrapper[4733]: I0202 15:36:24.801031 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d57375e-7e2f-4870-8c60-555921f94b5e-scripts\") pod \"manila-scheduler-2\" (UID: \"7d57375e-7e2f-4870-8c60-555921f94b5e\") " pod="manila-kuttl-tests/manila-scheduler-2" Feb 02 15:36:24 crc kubenswrapper[4733]: I0202 15:36:24.801075 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d57375e-7e2f-4870-8c60-555921f94b5e-config-data\") pod \"manila-scheduler-2\" (UID: \"7d57375e-7e2f-4870-8c60-555921f94b5e\") " pod="manila-kuttl-tests/manila-scheduler-2" Feb 02 15:36:24 crc kubenswrapper[4733]: I0202 15:36:24.801136 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7d57375e-7e2f-4870-8c60-555921f94b5e-etc-machine-id\") pod \"manila-scheduler-2\" (UID: \"7d57375e-7e2f-4870-8c60-555921f94b5e\") " pod="manila-kuttl-tests/manila-scheduler-2" Feb 02 15:36:24 crc kubenswrapper[4733]: I0202 15:36:24.808693 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d57375e-7e2f-4870-8c60-555921f94b5e-scripts\") pod \"manila-scheduler-2\" (UID: \"7d57375e-7e2f-4870-8c60-555921f94b5e\") " pod="manila-kuttl-tests/manila-scheduler-2" Feb 02 15:36:24 crc kubenswrapper[4733]: I0202 15:36:24.809382 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7d57375e-7e2f-4870-8c60-555921f94b5e-config-data-custom\") pod \"manila-scheduler-2\" (UID: \"7d57375e-7e2f-4870-8c60-555921f94b5e\") " pod="manila-kuttl-tests/manila-scheduler-2" Feb 02 15:36:24 crc kubenswrapper[4733]: I0202 15:36:24.809899 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d57375e-7e2f-4870-8c60-555921f94b5e-config-data\") pod \"manila-scheduler-2\" (UID: \"7d57375e-7e2f-4870-8c60-555921f94b5e\") " pod="manila-kuttl-tests/manila-scheduler-2" Feb 02 15:36:24 crc kubenswrapper[4733]: I0202 15:36:24.819122 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnn56\" (UniqueName: \"kubernetes.io/projected/7d57375e-7e2f-4870-8c60-555921f94b5e-kube-api-access-fnn56\") pod \"manila-scheduler-2\" (UID: \"7d57375e-7e2f-4870-8c60-555921f94b5e\") " pod="manila-kuttl-tests/manila-scheduler-2" Feb 02 15:36:24 crc kubenswrapper[4733]: I0202 15:36:24.921462 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-scheduler-2" Feb 02 15:36:25 crc kubenswrapper[4733]: I0202 15:36:25.407229 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/manila-scheduler-2"] Feb 02 15:36:26 crc kubenswrapper[4733]: I0202 15:36:26.216319 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-scheduler-2" event={"ID":"7d57375e-7e2f-4870-8c60-555921f94b5e","Type":"ContainerStarted","Data":"08fe607bae51dac681c85d8b2d9d672ff5358eacb16223407be6cb259e100696"} Feb 02 15:36:26 crc kubenswrapper[4733]: I0202 15:36:26.216658 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-scheduler-2" event={"ID":"7d57375e-7e2f-4870-8c60-555921f94b5e","Type":"ContainerStarted","Data":"7750f81528ff63574d44601336d185bd84fda8feb232933006a0f6e1e3724ce9"} Feb 02 15:36:27 crc kubenswrapper[4733]: I0202 15:36:27.229763 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-scheduler-2" event={"ID":"7d57375e-7e2f-4870-8c60-555921f94b5e","Type":"ContainerStarted","Data":"2e4215c5ac05a6aecd6ea68d6bed338b4de74d67b17bf07cc58e28e72f8f5b0a"} Feb 02 15:36:27 crc kubenswrapper[4733]: I0202 15:36:27.252756 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="manila-kuttl-tests/manila-scheduler-2" podStartSLOduration=3.252724562 podStartE2EDuration="3.252724562s" podCreationTimestamp="2026-02-02 15:36:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:36:27.247863144 +0000 UTC m=+1270.699324542" watchObservedRunningTime="2026-02-02 15:36:27.252724562 +0000 UTC m=+1270.704185980" Feb 02 15:36:34 crc kubenswrapper[4733]: I0202 15:36:34.922354 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="manila-kuttl-tests/manila-scheduler-2" Feb 02 15:36:34 crc kubenswrapper[4733]: I0202 15:36:34.987268 4733 patch_prober.go:28] interesting pod/machine-config-daemon-zgsv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 15:36:34 crc kubenswrapper[4733]: I0202 15:36:34.987357 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 15:36:34 crc kubenswrapper[4733]: I0202 15:36:34.987428 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" Feb 02 15:36:34 crc kubenswrapper[4733]: I0202 15:36:34.988304 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"26cec88244c10ffe50680f654ac54834881498bea583f15ab3590dda3791c1bf"} pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 15:36:34 crc kubenswrapper[4733]: I0202 15:36:34.988512 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" containerName="machine-config-daemon" containerID="cri-o://26cec88244c10ffe50680f654ac54834881498bea583f15ab3590dda3791c1bf" gracePeriod=600 Feb 02 15:36:35 crc kubenswrapper[4733]: I0202 15:36:35.304192 4733 generic.go:334] "Generic (PLEG): container finished" podID="83d0e810-ea5c-47aa-947b-7865b045bf94" containerID="26cec88244c10ffe50680f654ac54834881498bea583f15ab3590dda3791c1bf" exitCode=0 Feb 02 15:36:35 crc kubenswrapper[4733]: I0202 15:36:35.304351 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" event={"ID":"83d0e810-ea5c-47aa-947b-7865b045bf94","Type":"ContainerDied","Data":"26cec88244c10ffe50680f654ac54834881498bea583f15ab3590dda3791c1bf"} Feb 02 15:36:35 crc kubenswrapper[4733]: I0202 15:36:35.304728 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" event={"ID":"83d0e810-ea5c-47aa-947b-7865b045bf94","Type":"ContainerStarted","Data":"9ef5e0247a233902273cc6edac5d4e35e73d0f3d5ded3fb8bb539db06e43f773"} Feb 02 15:36:35 crc kubenswrapper[4733]: I0202 15:36:35.304769 4733 scope.go:117] "RemoveContainer" containerID="b050108ccfab6b78f46eeabfccf3d93e55ea50151667d0fea583d0f4073f7662" Feb 02 15:36:46 crc kubenswrapper[4733]: I0202 15:36:46.471316 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="manila-kuttl-tests/manila-scheduler-2" Feb 02 15:36:48 crc kubenswrapper[4733]: I0202 15:36:48.129365 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/manila-db-sync-fr7nc"] Feb 02 15:36:48 crc kubenswrapper[4733]: I0202 15:36:48.163258 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["manila-kuttl-tests/manila-db-sync-fr7nc"] Feb 02 15:36:48 crc kubenswrapper[4733]: I0202 15:36:48.172067 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/manila-share-share0-0"] Feb 02 15:36:48 crc kubenswrapper[4733]: I0202 15:36:48.172705 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="manila-kuttl-tests/manila-share-share0-0" podUID="c5853503-348a-4386-9fec-cfa5b11b7501" containerName="manila-share" containerID="cri-o://c8d9156f304d30a3ef84f3b663ebb0f580e2927d42d46e4cde55497b6111d850" gracePeriod=30 Feb 02 15:36:48 crc kubenswrapper[4733]: I0202 15:36:48.173216 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="manila-kuttl-tests/manila-share-share0-0" podUID="c5853503-348a-4386-9fec-cfa5b11b7501" containerName="probe" containerID="cri-o://116efcfe9abd31a050f2d88c205ff735cb117b043fea831b68e74f6b6fc08b1f" gracePeriod=30 Feb 02 15:36:48 crc kubenswrapper[4733]: I0202 15:36:48.187903 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/manila-scheduler-0"] Feb 02 15:36:48 crc kubenswrapper[4733]: I0202 15:36:48.188145 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="manila-kuttl-tests/manila-scheduler-0" podUID="d04d46e8-50b1-47ab-ac62-a7f446949411" containerName="manila-scheduler" containerID="cri-o://df75409bdb9d1900e091e8eac282b55cecc19bc761f50fd523a5af5a3b567d13" gracePeriod=30 Feb 02 15:36:48 crc kubenswrapper[4733]: I0202 15:36:48.188499 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="manila-kuttl-tests/manila-scheduler-0" podUID="d04d46e8-50b1-47ab-ac62-a7f446949411" containerName="probe" containerID="cri-o://be664a164e3540eacbc8f61c59d2f24b886ed58c44ac7a162eb05192733017a5" gracePeriod=30 Feb 02 15:36:48 crc kubenswrapper[4733]: I0202 15:36:48.197615 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/manila-scheduler-2"] Feb 02 15:36:48 crc kubenswrapper[4733]: I0202 15:36:48.197817 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="manila-kuttl-tests/manila-scheduler-2" podUID="7d57375e-7e2f-4870-8c60-555921f94b5e" containerName="manila-scheduler" containerID="cri-o://08fe607bae51dac681c85d8b2d9d672ff5358eacb16223407be6cb259e100696" gracePeriod=30 Feb 02 15:36:48 crc kubenswrapper[4733]: I0202 15:36:48.197936 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="manila-kuttl-tests/manila-scheduler-2" podUID="7d57375e-7e2f-4870-8c60-555921f94b5e" containerName="probe" containerID="cri-o://2e4215c5ac05a6aecd6ea68d6bed338b4de74d67b17bf07cc58e28e72f8f5b0a" gracePeriod=30 Feb 02 15:36:48 crc kubenswrapper[4733]: I0202 15:36:48.207402 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/manila-scheduler-1"] Feb 02 15:36:48 crc kubenswrapper[4733]: I0202 15:36:48.208614 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="manila-kuttl-tests/manila-scheduler-1" podUID="78bb94e9-3aad-4917-8a42-105b948191a6" containerName="probe" containerID="cri-o://c7f431f57e22c98695fc1863dff3e6d6c1ca0c21042e82a1b590b47d616d2bd7" gracePeriod=30 Feb 02 15:36:48 crc kubenswrapper[4733]: I0202 15:36:48.212040 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="manila-kuttl-tests/manila-scheduler-1" podUID="78bb94e9-3aad-4917-8a42-105b948191a6" containerName="manila-scheduler" containerID="cri-o://2ac63b9e1e48ebf555f89cf6b525c9b5d7b845fff893abba074691caf5188aaa" gracePeriod=30 Feb 02 15:36:48 crc kubenswrapper[4733]: I0202 15:36:48.224640 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["manila-kuttl-tests/manila6c72-account-delete-qcx45"] Feb 02 15:36:48 crc kubenswrapper[4733]: I0202 15:36:48.225507 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila6c72-account-delete-qcx45" Feb 02 15:36:48 crc kubenswrapper[4733]: I0202 15:36:48.249664 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/manila-api-0"] Feb 02 15:36:48 crc kubenswrapper[4733]: I0202 15:36:48.256231 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="manila-kuttl-tests/manila-api-0" podUID="b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a" containerName="manila-api-log" containerID="cri-o://a180a6fc32410e33617fc26a50c345f446bcc4a648a07a9f2d7b522c877aa0d1" gracePeriod=30 Feb 02 15:36:48 crc kubenswrapper[4733]: I0202 15:36:48.256382 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="manila-kuttl-tests/manila-api-0" podUID="b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a" containerName="manila-api" containerID="cri-o://a68fb69fc8aa87f439ba650174df8155e78d902fa293b457e31c20c2112e9bb7" gracePeriod=30 Feb 02 15:36:48 crc kubenswrapper[4733]: I0202 15:36:48.283236 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/manila6c72-account-delete-qcx45"] Feb 02 15:36:48 crc kubenswrapper[4733]: I0202 15:36:48.289013 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cth5w\" (UniqueName: \"kubernetes.io/projected/f0fe5bb8-57d0-401a-81d5-5e7fc9b967f7-kube-api-access-cth5w\") pod \"manila6c72-account-delete-qcx45\" (UID: \"f0fe5bb8-57d0-401a-81d5-5e7fc9b967f7\") " pod="manila-kuttl-tests/manila6c72-account-delete-qcx45" Feb 02 15:36:48 crc kubenswrapper[4733]: I0202 15:36:48.289140 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0fe5bb8-57d0-401a-81d5-5e7fc9b967f7-operator-scripts\") pod \"manila6c72-account-delete-qcx45\" (UID: \"f0fe5bb8-57d0-401a-81d5-5e7fc9b967f7\") " pod="manila-kuttl-tests/manila6c72-account-delete-qcx45" Feb 02 15:36:48 crc kubenswrapper[4733]: I0202 15:36:48.391078 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cth5w\" (UniqueName: \"kubernetes.io/projected/f0fe5bb8-57d0-401a-81d5-5e7fc9b967f7-kube-api-access-cth5w\") pod \"manila6c72-account-delete-qcx45\" (UID: \"f0fe5bb8-57d0-401a-81d5-5e7fc9b967f7\") " pod="manila-kuttl-tests/manila6c72-account-delete-qcx45" Feb 02 15:36:48 crc kubenswrapper[4733]: I0202 15:36:48.391133 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0fe5bb8-57d0-401a-81d5-5e7fc9b967f7-operator-scripts\") pod \"manila6c72-account-delete-qcx45\" (UID: \"f0fe5bb8-57d0-401a-81d5-5e7fc9b967f7\") " pod="manila-kuttl-tests/manila6c72-account-delete-qcx45" Feb 02 15:36:48 crc kubenswrapper[4733]: I0202 15:36:48.391957 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0fe5bb8-57d0-401a-81d5-5e7fc9b967f7-operator-scripts\") pod \"manila6c72-account-delete-qcx45\" (UID: \"f0fe5bb8-57d0-401a-81d5-5e7fc9b967f7\") " pod="manila-kuttl-tests/manila6c72-account-delete-qcx45" Feb 02 15:36:48 crc kubenswrapper[4733]: I0202 15:36:48.412926 4733 generic.go:334] "Generic (PLEG): container finished" podID="b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a" containerID="a180a6fc32410e33617fc26a50c345f446bcc4a648a07a9f2d7b522c877aa0d1" exitCode=143 Feb 02 15:36:48 crc kubenswrapper[4733]: I0202 15:36:48.412970 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-api-0" event={"ID":"b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a","Type":"ContainerDied","Data":"a180a6fc32410e33617fc26a50c345f446bcc4a648a07a9f2d7b522c877aa0d1"} Feb 02 15:36:48 crc kubenswrapper[4733]: I0202 15:36:48.414053 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cth5w\" (UniqueName: \"kubernetes.io/projected/f0fe5bb8-57d0-401a-81d5-5e7fc9b967f7-kube-api-access-cth5w\") pod \"manila6c72-account-delete-qcx45\" (UID: \"f0fe5bb8-57d0-401a-81d5-5e7fc9b967f7\") " pod="manila-kuttl-tests/manila6c72-account-delete-qcx45" Feb 02 15:36:48 crc kubenswrapper[4733]: I0202 15:36:48.553699 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila6c72-account-delete-qcx45" Feb 02 15:36:49 crc kubenswrapper[4733]: I0202 15:36:49.039878 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:36:49 crc kubenswrapper[4733]: I0202 15:36:49.096400 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/manila6c72-account-delete-qcx45"] Feb 02 15:36:49 crc kubenswrapper[4733]: I0202 15:36:49.099815 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c5853503-348a-4386-9fec-cfa5b11b7501-etc-machine-id\") pod \"c5853503-348a-4386-9fec-cfa5b11b7501\" (UID: \"c5853503-348a-4386-9fec-cfa5b11b7501\") " Feb 02 15:36:49 crc kubenswrapper[4733]: I0202 15:36:49.099927 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c5853503-348a-4386-9fec-cfa5b11b7501-config-data-custom\") pod \"c5853503-348a-4386-9fec-cfa5b11b7501\" (UID: \"c5853503-348a-4386-9fec-cfa5b11b7501\") " Feb 02 15:36:49 crc kubenswrapper[4733]: I0202 15:36:49.099970 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/c5853503-348a-4386-9fec-cfa5b11b7501-var-lib-manila\") pod \"c5853503-348a-4386-9fec-cfa5b11b7501\" (UID: \"c5853503-348a-4386-9fec-cfa5b11b7501\") " Feb 02 15:36:49 crc kubenswrapper[4733]: I0202 15:36:49.100010 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c5853503-348a-4386-9fec-cfa5b11b7501-ceph\") pod \"c5853503-348a-4386-9fec-cfa5b11b7501\" (UID: \"c5853503-348a-4386-9fec-cfa5b11b7501\") " Feb 02 15:36:49 crc kubenswrapper[4733]: I0202 15:36:49.100048 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5853503-348a-4386-9fec-cfa5b11b7501-scripts\") pod \"c5853503-348a-4386-9fec-cfa5b11b7501\" (UID: \"c5853503-348a-4386-9fec-cfa5b11b7501\") " Feb 02 15:36:49 crc kubenswrapper[4733]: I0202 15:36:49.100088 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5853503-348a-4386-9fec-cfa5b11b7501-config-data\") pod \"c5853503-348a-4386-9fec-cfa5b11b7501\" (UID: \"c5853503-348a-4386-9fec-cfa5b11b7501\") " Feb 02 15:36:49 crc kubenswrapper[4733]: I0202 15:36:49.100118 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-th7nw\" (UniqueName: \"kubernetes.io/projected/c5853503-348a-4386-9fec-cfa5b11b7501-kube-api-access-th7nw\") pod \"c5853503-348a-4386-9fec-cfa5b11b7501\" (UID: \"c5853503-348a-4386-9fec-cfa5b11b7501\") " Feb 02 15:36:49 crc kubenswrapper[4733]: I0202 15:36:49.101297 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c5853503-348a-4386-9fec-cfa5b11b7501-var-lib-manila" (OuterVolumeSpecName: "var-lib-manila") pod "c5853503-348a-4386-9fec-cfa5b11b7501" (UID: "c5853503-348a-4386-9fec-cfa5b11b7501"). InnerVolumeSpecName "var-lib-manila". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 15:36:49 crc kubenswrapper[4733]: I0202 15:36:49.101373 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c5853503-348a-4386-9fec-cfa5b11b7501-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "c5853503-348a-4386-9fec-cfa5b11b7501" (UID: "c5853503-348a-4386-9fec-cfa5b11b7501"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 15:36:49 crc kubenswrapper[4733]: I0202 15:36:49.105751 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5853503-348a-4386-9fec-cfa5b11b7501-ceph" (OuterVolumeSpecName: "ceph") pod "c5853503-348a-4386-9fec-cfa5b11b7501" (UID: "c5853503-348a-4386-9fec-cfa5b11b7501"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:36:49 crc kubenswrapper[4733]: I0202 15:36:49.106547 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5853503-348a-4386-9fec-cfa5b11b7501-scripts" (OuterVolumeSpecName: "scripts") pod "c5853503-348a-4386-9fec-cfa5b11b7501" (UID: "c5853503-348a-4386-9fec-cfa5b11b7501"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:36:49 crc kubenswrapper[4733]: I0202 15:36:49.106753 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5853503-348a-4386-9fec-cfa5b11b7501-kube-api-access-th7nw" (OuterVolumeSpecName: "kube-api-access-th7nw") pod "c5853503-348a-4386-9fec-cfa5b11b7501" (UID: "c5853503-348a-4386-9fec-cfa5b11b7501"). InnerVolumeSpecName "kube-api-access-th7nw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:36:49 crc kubenswrapper[4733]: I0202 15:36:49.107193 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5853503-348a-4386-9fec-cfa5b11b7501-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "c5853503-348a-4386-9fec-cfa5b11b7501" (UID: "c5853503-348a-4386-9fec-cfa5b11b7501"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:36:49 crc kubenswrapper[4733]: I0202 15:36:49.171659 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5853503-348a-4386-9fec-cfa5b11b7501-config-data" (OuterVolumeSpecName: "config-data") pod "c5853503-348a-4386-9fec-cfa5b11b7501" (UID: "c5853503-348a-4386-9fec-cfa5b11b7501"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:36:49 crc kubenswrapper[4733]: I0202 15:36:49.202316 4733 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c5853503-348a-4386-9fec-cfa5b11b7501-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 02 15:36:49 crc kubenswrapper[4733]: I0202 15:36:49.202354 4733 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c5853503-348a-4386-9fec-cfa5b11b7501-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 02 15:36:49 crc kubenswrapper[4733]: I0202 15:36:49.202366 4733 reconciler_common.go:293] "Volume detached for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/c5853503-348a-4386-9fec-cfa5b11b7501-var-lib-manila\") on node \"crc\" DevicePath \"\"" Feb 02 15:36:49 crc kubenswrapper[4733]: I0202 15:36:49.202375 4733 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c5853503-348a-4386-9fec-cfa5b11b7501-ceph\") on node \"crc\" DevicePath \"\"" Feb 02 15:36:49 crc kubenswrapper[4733]: I0202 15:36:49.202386 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5853503-348a-4386-9fec-cfa5b11b7501-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 15:36:49 crc kubenswrapper[4733]: I0202 15:36:49.202394 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5853503-348a-4386-9fec-cfa5b11b7501-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 15:36:49 crc kubenswrapper[4733]: I0202 15:36:49.202403 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-th7nw\" (UniqueName: \"kubernetes.io/projected/c5853503-348a-4386-9fec-cfa5b11b7501-kube-api-access-th7nw\") on node \"crc\" DevicePath \"\"" Feb 02 15:36:49 crc kubenswrapper[4733]: I0202 15:36:49.262357 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb872024-d057-4eaa-8f93-0401ac466650" path="/var/lib/kubelet/pods/eb872024-d057-4eaa-8f93-0401ac466650/volumes" Feb 02 15:36:49 crc kubenswrapper[4733]: I0202 15:36:49.420628 4733 generic.go:334] "Generic (PLEG): container finished" podID="f0fe5bb8-57d0-401a-81d5-5e7fc9b967f7" containerID="aa3ebde02e5a5d6bc82ad8bf39facfc176dfcc69833c63a12ba398b12435f40e" exitCode=0 Feb 02 15:36:49 crc kubenswrapper[4733]: I0202 15:36:49.420703 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila6c72-account-delete-qcx45" event={"ID":"f0fe5bb8-57d0-401a-81d5-5e7fc9b967f7","Type":"ContainerDied","Data":"aa3ebde02e5a5d6bc82ad8bf39facfc176dfcc69833c63a12ba398b12435f40e"} Feb 02 15:36:49 crc kubenswrapper[4733]: I0202 15:36:49.420740 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila6c72-account-delete-qcx45" event={"ID":"f0fe5bb8-57d0-401a-81d5-5e7fc9b967f7","Type":"ContainerStarted","Data":"1f1ac651f99a7c137766f936f4dfad7b23580337194be98748abf3c836edf255"} Feb 02 15:36:49 crc kubenswrapper[4733]: I0202 15:36:49.422548 4733 generic.go:334] "Generic (PLEG): container finished" podID="7d57375e-7e2f-4870-8c60-555921f94b5e" containerID="2e4215c5ac05a6aecd6ea68d6bed338b4de74d67b17bf07cc58e28e72f8f5b0a" exitCode=0 Feb 02 15:36:49 crc kubenswrapper[4733]: I0202 15:36:49.422620 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-scheduler-2" event={"ID":"7d57375e-7e2f-4870-8c60-555921f94b5e","Type":"ContainerDied","Data":"2e4215c5ac05a6aecd6ea68d6bed338b4de74d67b17bf07cc58e28e72f8f5b0a"} Feb 02 15:36:49 crc kubenswrapper[4733]: I0202 15:36:49.424448 4733 generic.go:334] "Generic (PLEG): container finished" podID="c5853503-348a-4386-9fec-cfa5b11b7501" containerID="116efcfe9abd31a050f2d88c205ff735cb117b043fea831b68e74f6b6fc08b1f" exitCode=0 Feb 02 15:36:49 crc kubenswrapper[4733]: I0202 15:36:49.424480 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-share-share0-0" event={"ID":"c5853503-348a-4386-9fec-cfa5b11b7501","Type":"ContainerDied","Data":"116efcfe9abd31a050f2d88c205ff735cb117b043fea831b68e74f6b6fc08b1f"} Feb 02 15:36:49 crc kubenswrapper[4733]: I0202 15:36:49.424498 4733 generic.go:334] "Generic (PLEG): container finished" podID="c5853503-348a-4386-9fec-cfa5b11b7501" containerID="c8d9156f304d30a3ef84f3b663ebb0f580e2927d42d46e4cde55497b6111d850" exitCode=1 Feb 02 15:36:49 crc kubenswrapper[4733]: I0202 15:36:49.424527 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-share-share0-0" event={"ID":"c5853503-348a-4386-9fec-cfa5b11b7501","Type":"ContainerDied","Data":"c8d9156f304d30a3ef84f3b663ebb0f580e2927d42d46e4cde55497b6111d850"} Feb 02 15:36:49 crc kubenswrapper[4733]: I0202 15:36:49.424538 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-share-share0-0" event={"ID":"c5853503-348a-4386-9fec-cfa5b11b7501","Type":"ContainerDied","Data":"c7b4d5086bca48301da0fee5eb1dac5a6b53ed5f8e8d8d2838325c643016ab25"} Feb 02 15:36:49 crc kubenswrapper[4733]: I0202 15:36:49.424663 4733 scope.go:117] "RemoveContainer" containerID="116efcfe9abd31a050f2d88c205ff735cb117b043fea831b68e74f6b6fc08b1f" Feb 02 15:36:49 crc kubenswrapper[4733]: I0202 15:36:49.424770 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:36:49 crc kubenswrapper[4733]: I0202 15:36:49.426828 4733 generic.go:334] "Generic (PLEG): container finished" podID="d04d46e8-50b1-47ab-ac62-a7f446949411" containerID="be664a164e3540eacbc8f61c59d2f24b886ed58c44ac7a162eb05192733017a5" exitCode=0 Feb 02 15:36:49 crc kubenswrapper[4733]: I0202 15:36:49.426878 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-scheduler-0" event={"ID":"d04d46e8-50b1-47ab-ac62-a7f446949411","Type":"ContainerDied","Data":"be664a164e3540eacbc8f61c59d2f24b886ed58c44ac7a162eb05192733017a5"} Feb 02 15:36:49 crc kubenswrapper[4733]: I0202 15:36:49.428657 4733 generic.go:334] "Generic (PLEG): container finished" podID="78bb94e9-3aad-4917-8a42-105b948191a6" containerID="c7f431f57e22c98695fc1863dff3e6d6c1ca0c21042e82a1b590b47d616d2bd7" exitCode=0 Feb 02 15:36:49 crc kubenswrapper[4733]: I0202 15:36:49.428684 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-scheduler-1" event={"ID":"78bb94e9-3aad-4917-8a42-105b948191a6","Type":"ContainerDied","Data":"c7f431f57e22c98695fc1863dff3e6d6c1ca0c21042e82a1b590b47d616d2bd7"} Feb 02 15:36:49 crc kubenswrapper[4733]: I0202 15:36:49.498032 4733 scope.go:117] "RemoveContainer" containerID="c8d9156f304d30a3ef84f3b663ebb0f580e2927d42d46e4cde55497b6111d850" Feb 02 15:36:49 crc kubenswrapper[4733]: I0202 15:36:49.504938 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/manila-share-share0-0"] Feb 02 15:36:49 crc kubenswrapper[4733]: I0202 15:36:49.521785 4733 scope.go:117] "RemoveContainer" containerID="116efcfe9abd31a050f2d88c205ff735cb117b043fea831b68e74f6b6fc08b1f" Feb 02 15:36:49 crc kubenswrapper[4733]: E0202 15:36:49.522581 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"116efcfe9abd31a050f2d88c205ff735cb117b043fea831b68e74f6b6fc08b1f\": container with ID starting with 116efcfe9abd31a050f2d88c205ff735cb117b043fea831b68e74f6b6fc08b1f not found: ID does not exist" containerID="116efcfe9abd31a050f2d88c205ff735cb117b043fea831b68e74f6b6fc08b1f" Feb 02 15:36:49 crc kubenswrapper[4733]: I0202 15:36:49.522623 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"116efcfe9abd31a050f2d88c205ff735cb117b043fea831b68e74f6b6fc08b1f"} err="failed to get container status \"116efcfe9abd31a050f2d88c205ff735cb117b043fea831b68e74f6b6fc08b1f\": rpc error: code = NotFound desc = could not find container \"116efcfe9abd31a050f2d88c205ff735cb117b043fea831b68e74f6b6fc08b1f\": container with ID starting with 116efcfe9abd31a050f2d88c205ff735cb117b043fea831b68e74f6b6fc08b1f not found: ID does not exist" Feb 02 15:36:49 crc kubenswrapper[4733]: I0202 15:36:49.522647 4733 scope.go:117] "RemoveContainer" containerID="c8d9156f304d30a3ef84f3b663ebb0f580e2927d42d46e4cde55497b6111d850" Feb 02 15:36:49 crc kubenswrapper[4733]: E0202 15:36:49.525663 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8d9156f304d30a3ef84f3b663ebb0f580e2927d42d46e4cde55497b6111d850\": container with ID starting with c8d9156f304d30a3ef84f3b663ebb0f580e2927d42d46e4cde55497b6111d850 not found: ID does not exist" containerID="c8d9156f304d30a3ef84f3b663ebb0f580e2927d42d46e4cde55497b6111d850" Feb 02 15:36:49 crc kubenswrapper[4733]: I0202 15:36:49.525694 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8d9156f304d30a3ef84f3b663ebb0f580e2927d42d46e4cde55497b6111d850"} err="failed to get container status \"c8d9156f304d30a3ef84f3b663ebb0f580e2927d42d46e4cde55497b6111d850\": rpc error: code = NotFound desc = could not find container \"c8d9156f304d30a3ef84f3b663ebb0f580e2927d42d46e4cde55497b6111d850\": container with ID starting with c8d9156f304d30a3ef84f3b663ebb0f580e2927d42d46e4cde55497b6111d850 not found: ID does not exist" Feb 02 15:36:49 crc kubenswrapper[4733]: I0202 15:36:49.525713 4733 scope.go:117] "RemoveContainer" containerID="116efcfe9abd31a050f2d88c205ff735cb117b043fea831b68e74f6b6fc08b1f" Feb 02 15:36:49 crc kubenswrapper[4733]: I0202 15:36:49.525951 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"116efcfe9abd31a050f2d88c205ff735cb117b043fea831b68e74f6b6fc08b1f"} err="failed to get container status \"116efcfe9abd31a050f2d88c205ff735cb117b043fea831b68e74f6b6fc08b1f\": rpc error: code = NotFound desc = could not find container \"116efcfe9abd31a050f2d88c205ff735cb117b043fea831b68e74f6b6fc08b1f\": container with ID starting with 116efcfe9abd31a050f2d88c205ff735cb117b043fea831b68e74f6b6fc08b1f not found: ID does not exist" Feb 02 15:36:49 crc kubenswrapper[4733]: I0202 15:36:49.525976 4733 scope.go:117] "RemoveContainer" containerID="c8d9156f304d30a3ef84f3b663ebb0f580e2927d42d46e4cde55497b6111d850" Feb 02 15:36:49 crc kubenswrapper[4733]: I0202 15:36:49.526370 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8d9156f304d30a3ef84f3b663ebb0f580e2927d42d46e4cde55497b6111d850"} err="failed to get container status \"c8d9156f304d30a3ef84f3b663ebb0f580e2927d42d46e4cde55497b6111d850\": rpc error: code = NotFound desc = could not find container \"c8d9156f304d30a3ef84f3b663ebb0f580e2927d42d46e4cde55497b6111d850\": container with ID starting with c8d9156f304d30a3ef84f3b663ebb0f580e2927d42d46e4cde55497b6111d850 not found: ID does not exist" Feb 02 15:36:49 crc kubenswrapper[4733]: I0202 15:36:49.527579 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["manila-kuttl-tests/manila-share-share0-0"] Feb 02 15:36:50 crc kubenswrapper[4733]: I0202 15:36:50.652069 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila6c72-account-delete-qcx45" Feb 02 15:36:50 crc kubenswrapper[4733]: I0202 15:36:50.722119 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0fe5bb8-57d0-401a-81d5-5e7fc9b967f7-operator-scripts\") pod \"f0fe5bb8-57d0-401a-81d5-5e7fc9b967f7\" (UID: \"f0fe5bb8-57d0-401a-81d5-5e7fc9b967f7\") " Feb 02 15:36:50 crc kubenswrapper[4733]: I0202 15:36:50.722234 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cth5w\" (UniqueName: \"kubernetes.io/projected/f0fe5bb8-57d0-401a-81d5-5e7fc9b967f7-kube-api-access-cth5w\") pod \"f0fe5bb8-57d0-401a-81d5-5e7fc9b967f7\" (UID: \"f0fe5bb8-57d0-401a-81d5-5e7fc9b967f7\") " Feb 02 15:36:50 crc kubenswrapper[4733]: I0202 15:36:50.722847 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0fe5bb8-57d0-401a-81d5-5e7fc9b967f7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f0fe5bb8-57d0-401a-81d5-5e7fc9b967f7" (UID: "f0fe5bb8-57d0-401a-81d5-5e7fc9b967f7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:36:50 crc kubenswrapper[4733]: I0202 15:36:50.726651 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0fe5bb8-57d0-401a-81d5-5e7fc9b967f7-kube-api-access-cth5w" (OuterVolumeSpecName: "kube-api-access-cth5w") pod "f0fe5bb8-57d0-401a-81d5-5e7fc9b967f7" (UID: "f0fe5bb8-57d0-401a-81d5-5e7fc9b967f7"). InnerVolumeSpecName "kube-api-access-cth5w". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:36:50 crc kubenswrapper[4733]: I0202 15:36:50.823941 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0fe5bb8-57d0-401a-81d5-5e7fc9b967f7-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 15:36:50 crc kubenswrapper[4733]: I0202 15:36:50.823984 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cth5w\" (UniqueName: \"kubernetes.io/projected/f0fe5bb8-57d0-401a-81d5-5e7fc9b967f7-kube-api-access-cth5w\") on node \"crc\" DevicePath \"\"" Feb 02 15:36:51 crc kubenswrapper[4733]: I0202 15:36:51.266776 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5853503-348a-4386-9fec-cfa5b11b7501" path="/var/lib/kubelet/pods/c5853503-348a-4386-9fec-cfa5b11b7501/volumes" Feb 02 15:36:51 crc kubenswrapper[4733]: I0202 15:36:51.406538 4733 prober.go:107] "Probe failed" probeType="Readiness" pod="manila-kuttl-tests/manila-api-0" podUID="b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a" containerName="manila-api" probeResult="failure" output="Get \"http://10.217.0.94:8786/healthcheck\": read tcp 10.217.0.2:59474->10.217.0.94:8786: read: connection reset by peer" Feb 02 15:36:51 crc kubenswrapper[4733]: I0202 15:36:51.444004 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila6c72-account-delete-qcx45" event={"ID":"f0fe5bb8-57d0-401a-81d5-5e7fc9b967f7","Type":"ContainerDied","Data":"1f1ac651f99a7c137766f936f4dfad7b23580337194be98748abf3c836edf255"} Feb 02 15:36:51 crc kubenswrapper[4733]: I0202 15:36:51.444044 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1f1ac651f99a7c137766f936f4dfad7b23580337194be98748abf3c836edf255" Feb 02 15:36:51 crc kubenswrapper[4733]: I0202 15:36:51.444098 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila6c72-account-delete-qcx45" Feb 02 15:36:51 crc kubenswrapper[4733]: I0202 15:36:51.804106 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-api-0" Feb 02 15:36:51 crc kubenswrapper[4733]: I0202 15:36:51.837950 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a-config-data\") pod \"b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a\" (UID: \"b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a\") " Feb 02 15:36:51 crc kubenswrapper[4733]: I0202 15:36:51.838284 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a-etc-machine-id\") pod \"b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a\" (UID: \"b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a\") " Feb 02 15:36:51 crc kubenswrapper[4733]: I0202 15:36:51.838407 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a" (UID: "b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 15:36:51 crc kubenswrapper[4733]: I0202 15:36:51.838420 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a-scripts\") pod \"b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a\" (UID: \"b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a\") " Feb 02 15:36:51 crc kubenswrapper[4733]: I0202 15:36:51.838569 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a-logs\") pod \"b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a\" (UID: \"b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a\") " Feb 02 15:36:51 crc kubenswrapper[4733]: I0202 15:36:51.838607 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a-config-data-custom\") pod \"b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a\" (UID: \"b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a\") " Feb 02 15:36:51 crc kubenswrapper[4733]: I0202 15:36:51.838703 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g4d6d\" (UniqueName: \"kubernetes.io/projected/b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a-kube-api-access-g4d6d\") pod \"b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a\" (UID: \"b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a\") " Feb 02 15:36:51 crc kubenswrapper[4733]: I0202 15:36:51.839016 4733 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 02 15:36:51 crc kubenswrapper[4733]: I0202 15:36:51.839233 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a-logs" (OuterVolumeSpecName: "logs") pod "b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a" (UID: "b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:36:51 crc kubenswrapper[4733]: I0202 15:36:51.843192 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a-kube-api-access-g4d6d" (OuterVolumeSpecName: "kube-api-access-g4d6d") pod "b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a" (UID: "b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a"). InnerVolumeSpecName "kube-api-access-g4d6d". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:36:51 crc kubenswrapper[4733]: I0202 15:36:51.843342 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a-scripts" (OuterVolumeSpecName: "scripts") pod "b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a" (UID: "b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:36:51 crc kubenswrapper[4733]: I0202 15:36:51.856828 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a" (UID: "b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:36:51 crc kubenswrapper[4733]: I0202 15:36:51.878658 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a-config-data" (OuterVolumeSpecName: "config-data") pod "b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a" (UID: "b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:36:51 crc kubenswrapper[4733]: I0202 15:36:51.939978 4733 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a-logs\") on node \"crc\" DevicePath \"\"" Feb 02 15:36:51 crc kubenswrapper[4733]: I0202 15:36:51.940011 4733 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 02 15:36:51 crc kubenswrapper[4733]: I0202 15:36:51.940022 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g4d6d\" (UniqueName: \"kubernetes.io/projected/b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a-kube-api-access-g4d6d\") on node \"crc\" DevicePath \"\"" Feb 02 15:36:51 crc kubenswrapper[4733]: I0202 15:36:51.940030 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 15:36:51 crc kubenswrapper[4733]: I0202 15:36:51.940041 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 15:36:52 crc kubenswrapper[4733]: I0202 15:36:52.453682 4733 generic.go:334] "Generic (PLEG): container finished" podID="b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a" containerID="a68fb69fc8aa87f439ba650174df8155e78d902fa293b457e31c20c2112e9bb7" exitCode=0 Feb 02 15:36:52 crc kubenswrapper[4733]: I0202 15:36:52.453840 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-api-0" event={"ID":"b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a","Type":"ContainerDied","Data":"a68fb69fc8aa87f439ba650174df8155e78d902fa293b457e31c20c2112e9bb7"} Feb 02 15:36:52 crc kubenswrapper[4733]: I0202 15:36:52.454003 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-api-0" event={"ID":"b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a","Type":"ContainerDied","Data":"2db7ec4764867ecfd65a0119375eb4ad8b7a57046d012cf74333eb00a046f8c5"} Feb 02 15:36:52 crc kubenswrapper[4733]: I0202 15:36:52.454029 4733 scope.go:117] "RemoveContainer" containerID="a68fb69fc8aa87f439ba650174df8155e78d902fa293b457e31c20c2112e9bb7" Feb 02 15:36:52 crc kubenswrapper[4733]: I0202 15:36:52.453919 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-api-0" Feb 02 15:36:52 crc kubenswrapper[4733]: I0202 15:36:52.481154 4733 scope.go:117] "RemoveContainer" containerID="a180a6fc32410e33617fc26a50c345f446bcc4a648a07a9f2d7b522c877aa0d1" Feb 02 15:36:52 crc kubenswrapper[4733]: I0202 15:36:52.490963 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/manila-api-0"] Feb 02 15:36:52 crc kubenswrapper[4733]: I0202 15:36:52.498283 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["manila-kuttl-tests/manila-api-0"] Feb 02 15:36:52 crc kubenswrapper[4733]: I0202 15:36:52.508339 4733 scope.go:117] "RemoveContainer" containerID="a68fb69fc8aa87f439ba650174df8155e78d902fa293b457e31c20c2112e9bb7" Feb 02 15:36:52 crc kubenswrapper[4733]: E0202 15:36:52.508789 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a68fb69fc8aa87f439ba650174df8155e78d902fa293b457e31c20c2112e9bb7\": container with ID starting with a68fb69fc8aa87f439ba650174df8155e78d902fa293b457e31c20c2112e9bb7 not found: ID does not exist" containerID="a68fb69fc8aa87f439ba650174df8155e78d902fa293b457e31c20c2112e9bb7" Feb 02 15:36:52 crc kubenswrapper[4733]: I0202 15:36:52.508817 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a68fb69fc8aa87f439ba650174df8155e78d902fa293b457e31c20c2112e9bb7"} err="failed to get container status \"a68fb69fc8aa87f439ba650174df8155e78d902fa293b457e31c20c2112e9bb7\": rpc error: code = NotFound desc = could not find container \"a68fb69fc8aa87f439ba650174df8155e78d902fa293b457e31c20c2112e9bb7\": container with ID starting with a68fb69fc8aa87f439ba650174df8155e78d902fa293b457e31c20c2112e9bb7 not found: ID does not exist" Feb 02 15:36:52 crc kubenswrapper[4733]: I0202 15:36:52.508837 4733 scope.go:117] "RemoveContainer" containerID="a180a6fc32410e33617fc26a50c345f446bcc4a648a07a9f2d7b522c877aa0d1" Feb 02 15:36:52 crc kubenswrapper[4733]: E0202 15:36:52.509211 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a180a6fc32410e33617fc26a50c345f446bcc4a648a07a9f2d7b522c877aa0d1\": container with ID starting with a180a6fc32410e33617fc26a50c345f446bcc4a648a07a9f2d7b522c877aa0d1 not found: ID does not exist" containerID="a180a6fc32410e33617fc26a50c345f446bcc4a648a07a9f2d7b522c877aa0d1" Feb 02 15:36:52 crc kubenswrapper[4733]: I0202 15:36:52.509231 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a180a6fc32410e33617fc26a50c345f446bcc4a648a07a9f2d7b522c877aa0d1"} err="failed to get container status \"a180a6fc32410e33617fc26a50c345f446bcc4a648a07a9f2d7b522c877aa0d1\": rpc error: code = NotFound desc = could not find container \"a180a6fc32410e33617fc26a50c345f446bcc4a648a07a9f2d7b522c877aa0d1\": container with ID starting with a180a6fc32410e33617fc26a50c345f446bcc4a648a07a9f2d7b522c877aa0d1 not found: ID does not exist" Feb 02 15:36:53 crc kubenswrapper[4733]: I0202 15:36:53.267159 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a" path="/var/lib/kubelet/pods/b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a/volumes" Feb 02 15:36:53 crc kubenswrapper[4733]: I0202 15:36:53.272960 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/manila-db-create-zlfrz"] Feb 02 15:36:53 crc kubenswrapper[4733]: I0202 15:36:53.291418 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["manila-kuttl-tests/manila-db-create-zlfrz"] Feb 02 15:36:53 crc kubenswrapper[4733]: I0202 15:36:53.301221 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/manila6c72-account-delete-qcx45"] Feb 02 15:36:53 crc kubenswrapper[4733]: I0202 15:36:53.309258 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["manila-kuttl-tests/manila6c72-account-delete-qcx45"] Feb 02 15:36:53 crc kubenswrapper[4733]: I0202 15:36:53.319589 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/manila-6c72-account-create-update-fkts6"] Feb 02 15:36:53 crc kubenswrapper[4733]: I0202 15:36:53.328458 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["manila-kuttl-tests/manila-6c72-account-create-update-fkts6"] Feb 02 15:36:53 crc kubenswrapper[4733]: I0202 15:36:53.446828 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["manila-kuttl-tests/manila-db-create-vq79x"] Feb 02 15:36:53 crc kubenswrapper[4733]: E0202 15:36:53.447202 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5853503-348a-4386-9fec-cfa5b11b7501" containerName="probe" Feb 02 15:36:53 crc kubenswrapper[4733]: I0202 15:36:53.447223 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5853503-348a-4386-9fec-cfa5b11b7501" containerName="probe" Feb 02 15:36:53 crc kubenswrapper[4733]: E0202 15:36:53.447241 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0fe5bb8-57d0-401a-81d5-5e7fc9b967f7" containerName="mariadb-account-delete" Feb 02 15:36:53 crc kubenswrapper[4733]: I0202 15:36:53.447264 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0fe5bb8-57d0-401a-81d5-5e7fc9b967f7" containerName="mariadb-account-delete" Feb 02 15:36:53 crc kubenswrapper[4733]: E0202 15:36:53.447278 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5853503-348a-4386-9fec-cfa5b11b7501" containerName="manila-share" Feb 02 15:36:53 crc kubenswrapper[4733]: I0202 15:36:53.447286 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5853503-348a-4386-9fec-cfa5b11b7501" containerName="manila-share" Feb 02 15:36:53 crc kubenswrapper[4733]: E0202 15:36:53.447299 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a" containerName="manila-api-log" Feb 02 15:36:53 crc kubenswrapper[4733]: I0202 15:36:53.447306 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a" containerName="manila-api-log" Feb 02 15:36:53 crc kubenswrapper[4733]: E0202 15:36:53.447318 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a" containerName="manila-api" Feb 02 15:36:53 crc kubenswrapper[4733]: I0202 15:36:53.447325 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a" containerName="manila-api" Feb 02 15:36:53 crc kubenswrapper[4733]: I0202 15:36:53.447468 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5853503-348a-4386-9fec-cfa5b11b7501" containerName="manila-share" Feb 02 15:36:53 crc kubenswrapper[4733]: I0202 15:36:53.447480 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a" containerName="manila-api" Feb 02 15:36:53 crc kubenswrapper[4733]: I0202 15:36:53.447492 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0fe5bb8-57d0-401a-81d5-5e7fc9b967f7" containerName="mariadb-account-delete" Feb 02 15:36:53 crc kubenswrapper[4733]: I0202 15:36:53.447500 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5853503-348a-4386-9fec-cfa5b11b7501" containerName="probe" Feb 02 15:36:53 crc kubenswrapper[4733]: I0202 15:36:53.447511 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3b0d234-8b3a-4b7b-9a43-64ae80e55a7a" containerName="manila-api-log" Feb 02 15:36:53 crc kubenswrapper[4733]: I0202 15:36:53.448055 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-db-create-vq79x" Feb 02 15:36:53 crc kubenswrapper[4733]: I0202 15:36:53.464745 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/manila-db-create-vq79x"] Feb 02 15:36:53 crc kubenswrapper[4733]: I0202 15:36:53.471696 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["manila-kuttl-tests/manila-eb2d-account-create-update-vzhkk"] Feb 02 15:36:53 crc kubenswrapper[4733]: I0202 15:36:53.472708 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-eb2d-account-create-update-vzhkk" Feb 02 15:36:53 crc kubenswrapper[4733]: I0202 15:36:53.476041 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"manila-db-secret" Feb 02 15:36:53 crc kubenswrapper[4733]: I0202 15:36:53.488019 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/manila-eb2d-account-create-update-vzhkk"] Feb 02 15:36:53 crc kubenswrapper[4733]: I0202 15:36:53.568197 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2761761e-fb0e-43f8-8ad9-5835b5f7e619-operator-scripts\") pod \"manila-db-create-vq79x\" (UID: \"2761761e-fb0e-43f8-8ad9-5835b5f7e619\") " pod="manila-kuttl-tests/manila-db-create-vq79x" Feb 02 15:36:53 crc kubenswrapper[4733]: I0202 15:36:53.568272 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4w64\" (UniqueName: \"kubernetes.io/projected/2761761e-fb0e-43f8-8ad9-5835b5f7e619-kube-api-access-h4w64\") pod \"manila-db-create-vq79x\" (UID: \"2761761e-fb0e-43f8-8ad9-5835b5f7e619\") " pod="manila-kuttl-tests/manila-db-create-vq79x" Feb 02 15:36:53 crc kubenswrapper[4733]: I0202 15:36:53.669985 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2761761e-fb0e-43f8-8ad9-5835b5f7e619-operator-scripts\") pod \"manila-db-create-vq79x\" (UID: \"2761761e-fb0e-43f8-8ad9-5835b5f7e619\") " pod="manila-kuttl-tests/manila-db-create-vq79x" Feb 02 15:36:53 crc kubenswrapper[4733]: I0202 15:36:53.670049 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4w64\" (UniqueName: \"kubernetes.io/projected/2761761e-fb0e-43f8-8ad9-5835b5f7e619-kube-api-access-h4w64\") pod \"manila-db-create-vq79x\" (UID: \"2761761e-fb0e-43f8-8ad9-5835b5f7e619\") " pod="manila-kuttl-tests/manila-db-create-vq79x" Feb 02 15:36:53 crc kubenswrapper[4733]: I0202 15:36:53.670093 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrmxk\" (UniqueName: \"kubernetes.io/projected/594f5269-2f23-4b68-8339-3871e122a649-kube-api-access-xrmxk\") pod \"manila-eb2d-account-create-update-vzhkk\" (UID: \"594f5269-2f23-4b68-8339-3871e122a649\") " pod="manila-kuttl-tests/manila-eb2d-account-create-update-vzhkk" Feb 02 15:36:53 crc kubenswrapper[4733]: I0202 15:36:53.670145 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/594f5269-2f23-4b68-8339-3871e122a649-operator-scripts\") pod \"manila-eb2d-account-create-update-vzhkk\" (UID: \"594f5269-2f23-4b68-8339-3871e122a649\") " pod="manila-kuttl-tests/manila-eb2d-account-create-update-vzhkk" Feb 02 15:36:53 crc kubenswrapper[4733]: I0202 15:36:53.671317 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2761761e-fb0e-43f8-8ad9-5835b5f7e619-operator-scripts\") pod \"manila-db-create-vq79x\" (UID: \"2761761e-fb0e-43f8-8ad9-5835b5f7e619\") " pod="manila-kuttl-tests/manila-db-create-vq79x" Feb 02 15:36:53 crc kubenswrapper[4733]: I0202 15:36:53.700471 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4w64\" (UniqueName: \"kubernetes.io/projected/2761761e-fb0e-43f8-8ad9-5835b5f7e619-kube-api-access-h4w64\") pod \"manila-db-create-vq79x\" (UID: \"2761761e-fb0e-43f8-8ad9-5835b5f7e619\") " pod="manila-kuttl-tests/manila-db-create-vq79x" Feb 02 15:36:53 crc kubenswrapper[4733]: I0202 15:36:53.766357 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-db-create-vq79x" Feb 02 15:36:53 crc kubenswrapper[4733]: I0202 15:36:53.772283 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrmxk\" (UniqueName: \"kubernetes.io/projected/594f5269-2f23-4b68-8339-3871e122a649-kube-api-access-xrmxk\") pod \"manila-eb2d-account-create-update-vzhkk\" (UID: \"594f5269-2f23-4b68-8339-3871e122a649\") " pod="manila-kuttl-tests/manila-eb2d-account-create-update-vzhkk" Feb 02 15:36:53 crc kubenswrapper[4733]: I0202 15:36:53.772341 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/594f5269-2f23-4b68-8339-3871e122a649-operator-scripts\") pod \"manila-eb2d-account-create-update-vzhkk\" (UID: \"594f5269-2f23-4b68-8339-3871e122a649\") " pod="manila-kuttl-tests/manila-eb2d-account-create-update-vzhkk" Feb 02 15:36:53 crc kubenswrapper[4733]: I0202 15:36:53.773381 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/594f5269-2f23-4b68-8339-3871e122a649-operator-scripts\") pod \"manila-eb2d-account-create-update-vzhkk\" (UID: \"594f5269-2f23-4b68-8339-3871e122a649\") " pod="manila-kuttl-tests/manila-eb2d-account-create-update-vzhkk" Feb 02 15:36:53 crc kubenswrapper[4733]: I0202 15:36:53.803606 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrmxk\" (UniqueName: \"kubernetes.io/projected/594f5269-2f23-4b68-8339-3871e122a649-kube-api-access-xrmxk\") pod \"manila-eb2d-account-create-update-vzhkk\" (UID: \"594f5269-2f23-4b68-8339-3871e122a649\") " pod="manila-kuttl-tests/manila-eb2d-account-create-update-vzhkk" Feb 02 15:36:53 crc kubenswrapper[4733]: I0202 15:36:53.993116 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.044317 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-scheduler-2" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.070484 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-scheduler-1" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.101528 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-eb2d-account-create-update-vzhkk" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.176615 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d04d46e8-50b1-47ab-ac62-a7f446949411-etc-machine-id\") pod \"d04d46e8-50b1-47ab-ac62-a7f446949411\" (UID: \"d04d46e8-50b1-47ab-ac62-a7f446949411\") " Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.176661 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7d57375e-7e2f-4870-8c60-555921f94b5e-etc-machine-id\") pod \"7d57375e-7e2f-4870-8c60-555921f94b5e\" (UID: \"7d57375e-7e2f-4870-8c60-555921f94b5e\") " Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.176694 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78bb94e9-3aad-4917-8a42-105b948191a6-scripts\") pod \"78bb94e9-3aad-4917-8a42-105b948191a6\" (UID: \"78bb94e9-3aad-4917-8a42-105b948191a6\") " Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.176737 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78bb94e9-3aad-4917-8a42-105b948191a6-config-data\") pod \"78bb94e9-3aad-4917-8a42-105b948191a6\" (UID: \"78bb94e9-3aad-4917-8a42-105b948191a6\") " Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.176761 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-drzqn\" (UniqueName: \"kubernetes.io/projected/d04d46e8-50b1-47ab-ac62-a7f446949411-kube-api-access-drzqn\") pod \"d04d46e8-50b1-47ab-ac62-a7f446949411\" (UID: \"d04d46e8-50b1-47ab-ac62-a7f446949411\") " Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.176786 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fnn56\" (UniqueName: \"kubernetes.io/projected/7d57375e-7e2f-4870-8c60-555921f94b5e-kube-api-access-fnn56\") pod \"7d57375e-7e2f-4870-8c60-555921f94b5e\" (UID: \"7d57375e-7e2f-4870-8c60-555921f94b5e\") " Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.176801 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d57375e-7e2f-4870-8c60-555921f94b5e-config-data\") pod \"7d57375e-7e2f-4870-8c60-555921f94b5e\" (UID: \"7d57375e-7e2f-4870-8c60-555921f94b5e\") " Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.176821 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d57375e-7e2f-4870-8c60-555921f94b5e-scripts\") pod \"7d57375e-7e2f-4870-8c60-555921f94b5e\" (UID: \"7d57375e-7e2f-4870-8c60-555921f94b5e\") " Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.176840 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d04d46e8-50b1-47ab-ac62-a7f446949411-config-data\") pod \"d04d46e8-50b1-47ab-ac62-a7f446949411\" (UID: \"d04d46e8-50b1-47ab-ac62-a7f446949411\") " Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.176854 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/78bb94e9-3aad-4917-8a42-105b948191a6-config-data-custom\") pod \"78bb94e9-3aad-4917-8a42-105b948191a6\" (UID: \"78bb94e9-3aad-4917-8a42-105b948191a6\") " Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.176877 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fkv4z\" (UniqueName: \"kubernetes.io/projected/78bb94e9-3aad-4917-8a42-105b948191a6-kube-api-access-fkv4z\") pod \"78bb94e9-3aad-4917-8a42-105b948191a6\" (UID: \"78bb94e9-3aad-4917-8a42-105b948191a6\") " Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.176899 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d04d46e8-50b1-47ab-ac62-a7f446949411-scripts\") pod \"d04d46e8-50b1-47ab-ac62-a7f446949411\" (UID: \"d04d46e8-50b1-47ab-ac62-a7f446949411\") " Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.176916 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7d57375e-7e2f-4870-8c60-555921f94b5e-config-data-custom\") pod \"7d57375e-7e2f-4870-8c60-555921f94b5e\" (UID: \"7d57375e-7e2f-4870-8c60-555921f94b5e\") " Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.176934 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/78bb94e9-3aad-4917-8a42-105b948191a6-etc-machine-id\") pod \"78bb94e9-3aad-4917-8a42-105b948191a6\" (UID: \"78bb94e9-3aad-4917-8a42-105b948191a6\") " Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.176973 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d04d46e8-50b1-47ab-ac62-a7f446949411-config-data-custom\") pod \"d04d46e8-50b1-47ab-ac62-a7f446949411\" (UID: \"d04d46e8-50b1-47ab-ac62-a7f446949411\") " Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.180854 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d04d46e8-50b1-47ab-ac62-a7f446949411-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d04d46e8-50b1-47ab-ac62-a7f446949411" (UID: "d04d46e8-50b1-47ab-ac62-a7f446949411"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.182604 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78bb94e9-3aad-4917-8a42-105b948191a6-scripts" (OuterVolumeSpecName: "scripts") pod "78bb94e9-3aad-4917-8a42-105b948191a6" (UID: "78bb94e9-3aad-4917-8a42-105b948191a6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.182667 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d04d46e8-50b1-47ab-ac62-a7f446949411-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "d04d46e8-50b1-47ab-ac62-a7f446949411" (UID: "d04d46e8-50b1-47ab-ac62-a7f446949411"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.182696 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7d57375e-7e2f-4870-8c60-555921f94b5e-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "7d57375e-7e2f-4870-8c60-555921f94b5e" (UID: "7d57375e-7e2f-4870-8c60-555921f94b5e"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.184111 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d57375e-7e2f-4870-8c60-555921f94b5e-scripts" (OuterVolumeSpecName: "scripts") pod "7d57375e-7e2f-4870-8c60-555921f94b5e" (UID: "7d57375e-7e2f-4870-8c60-555921f94b5e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.184207 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d04d46e8-50b1-47ab-ac62-a7f446949411-kube-api-access-drzqn" (OuterVolumeSpecName: "kube-api-access-drzqn") pod "d04d46e8-50b1-47ab-ac62-a7f446949411" (UID: "d04d46e8-50b1-47ab-ac62-a7f446949411"). InnerVolumeSpecName "kube-api-access-drzqn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.185076 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d57375e-7e2f-4870-8c60-555921f94b5e-kube-api-access-fnn56" (OuterVolumeSpecName: "kube-api-access-fnn56") pod "7d57375e-7e2f-4870-8c60-555921f94b5e" (UID: "7d57375e-7e2f-4870-8c60-555921f94b5e"). InnerVolumeSpecName "kube-api-access-fnn56". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.185129 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/78bb94e9-3aad-4917-8a42-105b948191a6-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "78bb94e9-3aad-4917-8a42-105b948191a6" (UID: "78bb94e9-3aad-4917-8a42-105b948191a6"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.186799 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d04d46e8-50b1-47ab-ac62-a7f446949411-scripts" (OuterVolumeSpecName: "scripts") pod "d04d46e8-50b1-47ab-ac62-a7f446949411" (UID: "d04d46e8-50b1-47ab-ac62-a7f446949411"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.186884 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d57375e-7e2f-4870-8c60-555921f94b5e-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "7d57375e-7e2f-4870-8c60-555921f94b5e" (UID: "7d57375e-7e2f-4870-8c60-555921f94b5e"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.189360 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78bb94e9-3aad-4917-8a42-105b948191a6-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "78bb94e9-3aad-4917-8a42-105b948191a6" (UID: "78bb94e9-3aad-4917-8a42-105b948191a6"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.191690 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78bb94e9-3aad-4917-8a42-105b948191a6-kube-api-access-fkv4z" (OuterVolumeSpecName: "kube-api-access-fkv4z") pod "78bb94e9-3aad-4917-8a42-105b948191a6" (UID: "78bb94e9-3aad-4917-8a42-105b948191a6"). InnerVolumeSpecName "kube-api-access-fkv4z". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.247202 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d57375e-7e2f-4870-8c60-555921f94b5e-config-data" (OuterVolumeSpecName: "config-data") pod "7d57375e-7e2f-4870-8c60-555921f94b5e" (UID: "7d57375e-7e2f-4870-8c60-555921f94b5e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.268266 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78bb94e9-3aad-4917-8a42-105b948191a6-config-data" (OuterVolumeSpecName: "config-data") pod "78bb94e9-3aad-4917-8a42-105b948191a6" (UID: "78bb94e9-3aad-4917-8a42-105b948191a6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.271405 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d04d46e8-50b1-47ab-ac62-a7f446949411-config-data" (OuterVolumeSpecName: "config-data") pod "d04d46e8-50b1-47ab-ac62-a7f446949411" (UID: "d04d46e8-50b1-47ab-ac62-a7f446949411"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.278542 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d04d46e8-50b1-47ab-ac62-a7f446949411-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.278573 4733 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7d57375e-7e2f-4870-8c60-555921f94b5e-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.278588 4733 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/78bb94e9-3aad-4917-8a42-105b948191a6-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.278600 4733 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d04d46e8-50b1-47ab-ac62-a7f446949411-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.278611 4733 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d04d46e8-50b1-47ab-ac62-a7f446949411-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.278622 4733 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7d57375e-7e2f-4870-8c60-555921f94b5e-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.278632 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78bb94e9-3aad-4917-8a42-105b948191a6-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.278642 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78bb94e9-3aad-4917-8a42-105b948191a6-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.278653 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-drzqn\" (UniqueName: \"kubernetes.io/projected/d04d46e8-50b1-47ab-ac62-a7f446949411-kube-api-access-drzqn\") on node \"crc\" DevicePath \"\"" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.278665 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fnn56\" (UniqueName: \"kubernetes.io/projected/7d57375e-7e2f-4870-8c60-555921f94b5e-kube-api-access-fnn56\") on node \"crc\" DevicePath \"\"" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.278677 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d57375e-7e2f-4870-8c60-555921f94b5e-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.278689 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d57375e-7e2f-4870-8c60-555921f94b5e-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.278700 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d04d46e8-50b1-47ab-ac62-a7f446949411-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.278710 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fkv4z\" (UniqueName: \"kubernetes.io/projected/78bb94e9-3aad-4917-8a42-105b948191a6-kube-api-access-fkv4z\") on node \"crc\" DevicePath \"\"" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.278722 4733 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/78bb94e9-3aad-4917-8a42-105b948191a6-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.279875 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/manila-db-create-vq79x"] Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.472597 4733 generic.go:334] "Generic (PLEG): container finished" podID="d04d46e8-50b1-47ab-ac62-a7f446949411" containerID="df75409bdb9d1900e091e8eac282b55cecc19bc761f50fd523a5af5a3b567d13" exitCode=0 Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.472682 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-scheduler-0" event={"ID":"d04d46e8-50b1-47ab-ac62-a7f446949411","Type":"ContainerDied","Data":"df75409bdb9d1900e091e8eac282b55cecc19bc761f50fd523a5af5a3b567d13"} Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.472700 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.472995 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-scheduler-0" event={"ID":"d04d46e8-50b1-47ab-ac62-a7f446949411","Type":"ContainerDied","Data":"1bc04e3fe0d9ec9a8cf631d5103d6ae3b217ccd6ae514fe3fe4634171563531b"} Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.473026 4733 scope.go:117] "RemoveContainer" containerID="be664a164e3540eacbc8f61c59d2f24b886ed58c44ac7a162eb05192733017a5" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.474784 4733 generic.go:334] "Generic (PLEG): container finished" podID="78bb94e9-3aad-4917-8a42-105b948191a6" containerID="2ac63b9e1e48ebf555f89cf6b525c9b5d7b845fff893abba074691caf5188aaa" exitCode=0 Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.474839 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-scheduler-1" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.474845 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-scheduler-1" event={"ID":"78bb94e9-3aad-4917-8a42-105b948191a6","Type":"ContainerDied","Data":"2ac63b9e1e48ebf555f89cf6b525c9b5d7b845fff893abba074691caf5188aaa"} Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.474871 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-scheduler-1" event={"ID":"78bb94e9-3aad-4917-8a42-105b948191a6","Type":"ContainerDied","Data":"d0efc3f8c2c8ee1132fc513c873eca127db5a1816c15b4348a02bcd87af57b76"} Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.479401 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-db-create-vq79x" event={"ID":"2761761e-fb0e-43f8-8ad9-5835b5f7e619","Type":"ContainerStarted","Data":"4f7eae9add368d7a939b590643cc6b2b137344a34ce936373e07f02c65d9fd86"} Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.479432 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-db-create-vq79x" event={"ID":"2761761e-fb0e-43f8-8ad9-5835b5f7e619","Type":"ContainerStarted","Data":"ad66327fda7063e6ee4998f5da5622c7aef043cd0bcaddbf443b3c38952c0dd0"} Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.482817 4733 generic.go:334] "Generic (PLEG): container finished" podID="7d57375e-7e2f-4870-8c60-555921f94b5e" containerID="08fe607bae51dac681c85d8b2d9d672ff5358eacb16223407be6cb259e100696" exitCode=0 Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.482859 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-scheduler-2" event={"ID":"7d57375e-7e2f-4870-8c60-555921f94b5e","Type":"ContainerDied","Data":"08fe607bae51dac681c85d8b2d9d672ff5358eacb16223407be6cb259e100696"} Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.482883 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-scheduler-2" event={"ID":"7d57375e-7e2f-4870-8c60-555921f94b5e","Type":"ContainerDied","Data":"7750f81528ff63574d44601336d185bd84fda8feb232933006a0f6e1e3724ce9"} Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.482946 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-scheduler-2" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.497568 4733 scope.go:117] "RemoveContainer" containerID="df75409bdb9d1900e091e8eac282b55cecc19bc761f50fd523a5af5a3b567d13" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.506393 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/manila-eb2d-account-create-update-vzhkk"] Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.514493 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="manila-kuttl-tests/manila-db-create-vq79x" podStartSLOduration=1.514469949 podStartE2EDuration="1.514469949s" podCreationTimestamp="2026-02-02 15:36:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:36:54.509039388 +0000 UTC m=+1297.960500776" watchObservedRunningTime="2026-02-02 15:36:54.514469949 +0000 UTC m=+1297.965931327" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.533680 4733 scope.go:117] "RemoveContainer" containerID="be664a164e3540eacbc8f61c59d2f24b886ed58c44ac7a162eb05192733017a5" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.535543 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/manila-scheduler-0"] Feb 02 15:36:54 crc kubenswrapper[4733]: E0202 15:36:54.535839 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be664a164e3540eacbc8f61c59d2f24b886ed58c44ac7a162eb05192733017a5\": container with ID starting with be664a164e3540eacbc8f61c59d2f24b886ed58c44ac7a162eb05192733017a5 not found: ID does not exist" containerID="be664a164e3540eacbc8f61c59d2f24b886ed58c44ac7a162eb05192733017a5" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.535892 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be664a164e3540eacbc8f61c59d2f24b886ed58c44ac7a162eb05192733017a5"} err="failed to get container status \"be664a164e3540eacbc8f61c59d2f24b886ed58c44ac7a162eb05192733017a5\": rpc error: code = NotFound desc = could not find container \"be664a164e3540eacbc8f61c59d2f24b886ed58c44ac7a162eb05192733017a5\": container with ID starting with be664a164e3540eacbc8f61c59d2f24b886ed58c44ac7a162eb05192733017a5 not found: ID does not exist" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.535924 4733 scope.go:117] "RemoveContainer" containerID="df75409bdb9d1900e091e8eac282b55cecc19bc761f50fd523a5af5a3b567d13" Feb 02 15:36:54 crc kubenswrapper[4733]: E0202 15:36:54.536598 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df75409bdb9d1900e091e8eac282b55cecc19bc761f50fd523a5af5a3b567d13\": container with ID starting with df75409bdb9d1900e091e8eac282b55cecc19bc761f50fd523a5af5a3b567d13 not found: ID does not exist" containerID="df75409bdb9d1900e091e8eac282b55cecc19bc761f50fd523a5af5a3b567d13" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.536639 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df75409bdb9d1900e091e8eac282b55cecc19bc761f50fd523a5af5a3b567d13"} err="failed to get container status \"df75409bdb9d1900e091e8eac282b55cecc19bc761f50fd523a5af5a3b567d13\": rpc error: code = NotFound desc = could not find container \"df75409bdb9d1900e091e8eac282b55cecc19bc761f50fd523a5af5a3b567d13\": container with ID starting with df75409bdb9d1900e091e8eac282b55cecc19bc761f50fd523a5af5a3b567d13 not found: ID does not exist" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.536665 4733 scope.go:117] "RemoveContainer" containerID="c7f431f57e22c98695fc1863dff3e6d6c1ca0c21042e82a1b590b47d616d2bd7" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.546188 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["manila-kuttl-tests/manila-scheduler-0"] Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.553437 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/manila-scheduler-2"] Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.560503 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["manila-kuttl-tests/manila-scheduler-2"] Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.566202 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/manila-scheduler-1"] Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.570668 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["manila-kuttl-tests/manila-scheduler-1"] Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.579774 4733 scope.go:117] "RemoveContainer" containerID="2ac63b9e1e48ebf555f89cf6b525c9b5d7b845fff893abba074691caf5188aaa" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.606486 4733 scope.go:117] "RemoveContainer" containerID="c7f431f57e22c98695fc1863dff3e6d6c1ca0c21042e82a1b590b47d616d2bd7" Feb 02 15:36:54 crc kubenswrapper[4733]: E0202 15:36:54.607119 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7f431f57e22c98695fc1863dff3e6d6c1ca0c21042e82a1b590b47d616d2bd7\": container with ID starting with c7f431f57e22c98695fc1863dff3e6d6c1ca0c21042e82a1b590b47d616d2bd7 not found: ID does not exist" containerID="c7f431f57e22c98695fc1863dff3e6d6c1ca0c21042e82a1b590b47d616d2bd7" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.607148 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7f431f57e22c98695fc1863dff3e6d6c1ca0c21042e82a1b590b47d616d2bd7"} err="failed to get container status \"c7f431f57e22c98695fc1863dff3e6d6c1ca0c21042e82a1b590b47d616d2bd7\": rpc error: code = NotFound desc = could not find container \"c7f431f57e22c98695fc1863dff3e6d6c1ca0c21042e82a1b590b47d616d2bd7\": container with ID starting with c7f431f57e22c98695fc1863dff3e6d6c1ca0c21042e82a1b590b47d616d2bd7 not found: ID does not exist" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.607222 4733 scope.go:117] "RemoveContainer" containerID="2ac63b9e1e48ebf555f89cf6b525c9b5d7b845fff893abba074691caf5188aaa" Feb 02 15:36:54 crc kubenswrapper[4733]: E0202 15:36:54.608309 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ac63b9e1e48ebf555f89cf6b525c9b5d7b845fff893abba074691caf5188aaa\": container with ID starting with 2ac63b9e1e48ebf555f89cf6b525c9b5d7b845fff893abba074691caf5188aaa not found: ID does not exist" containerID="2ac63b9e1e48ebf555f89cf6b525c9b5d7b845fff893abba074691caf5188aaa" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.608344 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ac63b9e1e48ebf555f89cf6b525c9b5d7b845fff893abba074691caf5188aaa"} err="failed to get container status \"2ac63b9e1e48ebf555f89cf6b525c9b5d7b845fff893abba074691caf5188aaa\": rpc error: code = NotFound desc = could not find container \"2ac63b9e1e48ebf555f89cf6b525c9b5d7b845fff893abba074691caf5188aaa\": container with ID starting with 2ac63b9e1e48ebf555f89cf6b525c9b5d7b845fff893abba074691caf5188aaa not found: ID does not exist" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.608369 4733 scope.go:117] "RemoveContainer" containerID="2e4215c5ac05a6aecd6ea68d6bed338b4de74d67b17bf07cc58e28e72f8f5b0a" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.732929 4733 scope.go:117] "RemoveContainer" containerID="08fe607bae51dac681c85d8b2d9d672ff5358eacb16223407be6cb259e100696" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.753208 4733 scope.go:117] "RemoveContainer" containerID="2e4215c5ac05a6aecd6ea68d6bed338b4de74d67b17bf07cc58e28e72f8f5b0a" Feb 02 15:36:54 crc kubenswrapper[4733]: E0202 15:36:54.754651 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e4215c5ac05a6aecd6ea68d6bed338b4de74d67b17bf07cc58e28e72f8f5b0a\": container with ID starting with 2e4215c5ac05a6aecd6ea68d6bed338b4de74d67b17bf07cc58e28e72f8f5b0a not found: ID does not exist" containerID="2e4215c5ac05a6aecd6ea68d6bed338b4de74d67b17bf07cc58e28e72f8f5b0a" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.754701 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e4215c5ac05a6aecd6ea68d6bed338b4de74d67b17bf07cc58e28e72f8f5b0a"} err="failed to get container status \"2e4215c5ac05a6aecd6ea68d6bed338b4de74d67b17bf07cc58e28e72f8f5b0a\": rpc error: code = NotFound desc = could not find container \"2e4215c5ac05a6aecd6ea68d6bed338b4de74d67b17bf07cc58e28e72f8f5b0a\": container with ID starting with 2e4215c5ac05a6aecd6ea68d6bed338b4de74d67b17bf07cc58e28e72f8f5b0a not found: ID does not exist" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.754731 4733 scope.go:117] "RemoveContainer" containerID="08fe607bae51dac681c85d8b2d9d672ff5358eacb16223407be6cb259e100696" Feb 02 15:36:54 crc kubenswrapper[4733]: E0202 15:36:54.755277 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"08fe607bae51dac681c85d8b2d9d672ff5358eacb16223407be6cb259e100696\": container with ID starting with 08fe607bae51dac681c85d8b2d9d672ff5358eacb16223407be6cb259e100696 not found: ID does not exist" containerID="08fe607bae51dac681c85d8b2d9d672ff5358eacb16223407be6cb259e100696" Feb 02 15:36:54 crc kubenswrapper[4733]: I0202 15:36:54.755317 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08fe607bae51dac681c85d8b2d9d672ff5358eacb16223407be6cb259e100696"} err="failed to get container status \"08fe607bae51dac681c85d8b2d9d672ff5358eacb16223407be6cb259e100696\": rpc error: code = NotFound desc = could not find container \"08fe607bae51dac681c85d8b2d9d672ff5358eacb16223407be6cb259e100696\": container with ID starting with 08fe607bae51dac681c85d8b2d9d672ff5358eacb16223407be6cb259e100696 not found: ID does not exist" Feb 02 15:36:55 crc kubenswrapper[4733]: I0202 15:36:55.261893 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77b06765-7874-4e7c-a681-d5da283dae25" path="/var/lib/kubelet/pods/77b06765-7874-4e7c-a681-d5da283dae25/volumes" Feb 02 15:36:55 crc kubenswrapper[4733]: I0202 15:36:55.262515 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78bb94e9-3aad-4917-8a42-105b948191a6" path="/var/lib/kubelet/pods/78bb94e9-3aad-4917-8a42-105b948191a6/volumes" Feb 02 15:36:55 crc kubenswrapper[4733]: I0202 15:36:55.263029 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d57375e-7e2f-4870-8c60-555921f94b5e" path="/var/lib/kubelet/pods/7d57375e-7e2f-4870-8c60-555921f94b5e/volumes" Feb 02 15:36:55 crc kubenswrapper[4733]: I0202 15:36:55.263968 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae455916-7075-477b-a71f-04537587c8ac" path="/var/lib/kubelet/pods/ae455916-7075-477b-a71f-04537587c8ac/volumes" Feb 02 15:36:55 crc kubenswrapper[4733]: I0202 15:36:55.264451 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d04d46e8-50b1-47ab-ac62-a7f446949411" path="/var/lib/kubelet/pods/d04d46e8-50b1-47ab-ac62-a7f446949411/volumes" Feb 02 15:36:55 crc kubenswrapper[4733]: I0202 15:36:55.264970 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0fe5bb8-57d0-401a-81d5-5e7fc9b967f7" path="/var/lib/kubelet/pods/f0fe5bb8-57d0-401a-81d5-5e7fc9b967f7/volumes" Feb 02 15:36:55 crc kubenswrapper[4733]: I0202 15:36:55.493591 4733 generic.go:334] "Generic (PLEG): container finished" podID="2761761e-fb0e-43f8-8ad9-5835b5f7e619" containerID="4f7eae9add368d7a939b590643cc6b2b137344a34ce936373e07f02c65d9fd86" exitCode=0 Feb 02 15:36:55 crc kubenswrapper[4733]: I0202 15:36:55.493665 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-db-create-vq79x" event={"ID":"2761761e-fb0e-43f8-8ad9-5835b5f7e619","Type":"ContainerDied","Data":"4f7eae9add368d7a939b590643cc6b2b137344a34ce936373e07f02c65d9fd86"} Feb 02 15:36:55 crc kubenswrapper[4733]: I0202 15:36:55.500453 4733 generic.go:334] "Generic (PLEG): container finished" podID="594f5269-2f23-4b68-8339-3871e122a649" containerID="1093dfe86ecb94863ae3a05142c42aa595e42408f929d5bf60be16907a8a2fd3" exitCode=0 Feb 02 15:36:55 crc kubenswrapper[4733]: I0202 15:36:55.500526 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-eb2d-account-create-update-vzhkk" event={"ID":"594f5269-2f23-4b68-8339-3871e122a649","Type":"ContainerDied","Data":"1093dfe86ecb94863ae3a05142c42aa595e42408f929d5bf60be16907a8a2fd3"} Feb 02 15:36:55 crc kubenswrapper[4733]: I0202 15:36:55.500564 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-eb2d-account-create-update-vzhkk" event={"ID":"594f5269-2f23-4b68-8339-3871e122a649","Type":"ContainerStarted","Data":"0f01cd71fc63d06bc993bbb8a131ceccd07ff9ab5ac69aeb0982421a4e6c561b"} Feb 02 15:36:56 crc kubenswrapper[4733]: I0202 15:36:56.864245 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-eb2d-account-create-update-vzhkk" Feb 02 15:36:56 crc kubenswrapper[4733]: I0202 15:36:56.869042 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-db-create-vq79x" Feb 02 15:36:57 crc kubenswrapper[4733]: I0202 15:36:57.016597 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xrmxk\" (UniqueName: \"kubernetes.io/projected/594f5269-2f23-4b68-8339-3871e122a649-kube-api-access-xrmxk\") pod \"594f5269-2f23-4b68-8339-3871e122a649\" (UID: \"594f5269-2f23-4b68-8339-3871e122a649\") " Feb 02 15:36:57 crc kubenswrapper[4733]: I0202 15:36:57.016647 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/594f5269-2f23-4b68-8339-3871e122a649-operator-scripts\") pod \"594f5269-2f23-4b68-8339-3871e122a649\" (UID: \"594f5269-2f23-4b68-8339-3871e122a649\") " Feb 02 15:36:57 crc kubenswrapper[4733]: I0202 15:36:57.016790 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2761761e-fb0e-43f8-8ad9-5835b5f7e619-operator-scripts\") pod \"2761761e-fb0e-43f8-8ad9-5835b5f7e619\" (UID: \"2761761e-fb0e-43f8-8ad9-5835b5f7e619\") " Feb 02 15:36:57 crc kubenswrapper[4733]: I0202 15:36:57.016979 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h4w64\" (UniqueName: \"kubernetes.io/projected/2761761e-fb0e-43f8-8ad9-5835b5f7e619-kube-api-access-h4w64\") pod \"2761761e-fb0e-43f8-8ad9-5835b5f7e619\" (UID: \"2761761e-fb0e-43f8-8ad9-5835b5f7e619\") " Feb 02 15:36:57 crc kubenswrapper[4733]: I0202 15:36:57.018012 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/594f5269-2f23-4b68-8339-3871e122a649-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "594f5269-2f23-4b68-8339-3871e122a649" (UID: "594f5269-2f23-4b68-8339-3871e122a649"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:36:57 crc kubenswrapper[4733]: I0202 15:36:57.018013 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2761761e-fb0e-43f8-8ad9-5835b5f7e619-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2761761e-fb0e-43f8-8ad9-5835b5f7e619" (UID: "2761761e-fb0e-43f8-8ad9-5835b5f7e619"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:36:57 crc kubenswrapper[4733]: I0202 15:36:57.022771 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/594f5269-2f23-4b68-8339-3871e122a649-kube-api-access-xrmxk" (OuterVolumeSpecName: "kube-api-access-xrmxk") pod "594f5269-2f23-4b68-8339-3871e122a649" (UID: "594f5269-2f23-4b68-8339-3871e122a649"). InnerVolumeSpecName "kube-api-access-xrmxk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:36:57 crc kubenswrapper[4733]: I0202 15:36:57.022898 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2761761e-fb0e-43f8-8ad9-5835b5f7e619-kube-api-access-h4w64" (OuterVolumeSpecName: "kube-api-access-h4w64") pod "2761761e-fb0e-43f8-8ad9-5835b5f7e619" (UID: "2761761e-fb0e-43f8-8ad9-5835b5f7e619"). InnerVolumeSpecName "kube-api-access-h4w64". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:36:57 crc kubenswrapper[4733]: I0202 15:36:57.118937 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h4w64\" (UniqueName: \"kubernetes.io/projected/2761761e-fb0e-43f8-8ad9-5835b5f7e619-kube-api-access-h4w64\") on node \"crc\" DevicePath \"\"" Feb 02 15:36:57 crc kubenswrapper[4733]: I0202 15:36:57.118972 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xrmxk\" (UniqueName: \"kubernetes.io/projected/594f5269-2f23-4b68-8339-3871e122a649-kube-api-access-xrmxk\") on node \"crc\" DevicePath \"\"" Feb 02 15:36:57 crc kubenswrapper[4733]: I0202 15:36:57.118982 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/594f5269-2f23-4b68-8339-3871e122a649-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 15:36:57 crc kubenswrapper[4733]: I0202 15:36:57.118991 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2761761e-fb0e-43f8-8ad9-5835b5f7e619-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 15:36:57 crc kubenswrapper[4733]: I0202 15:36:57.513230 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-db-create-vq79x" event={"ID":"2761761e-fb0e-43f8-8ad9-5835b5f7e619","Type":"ContainerDied","Data":"ad66327fda7063e6ee4998f5da5622c7aef043cd0bcaddbf443b3c38952c0dd0"} Feb 02 15:36:57 crc kubenswrapper[4733]: I0202 15:36:57.513274 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ad66327fda7063e6ee4998f5da5622c7aef043cd0bcaddbf443b3c38952c0dd0" Feb 02 15:36:57 crc kubenswrapper[4733]: I0202 15:36:57.513527 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-db-create-vq79x" Feb 02 15:36:57 crc kubenswrapper[4733]: I0202 15:36:57.514746 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-eb2d-account-create-update-vzhkk" event={"ID":"594f5269-2f23-4b68-8339-3871e122a649","Type":"ContainerDied","Data":"0f01cd71fc63d06bc993bbb8a131ceccd07ff9ab5ac69aeb0982421a4e6c561b"} Feb 02 15:36:57 crc kubenswrapper[4733]: I0202 15:36:57.514905 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0f01cd71fc63d06bc993bbb8a131ceccd07ff9ab5ac69aeb0982421a4e6c561b" Feb 02 15:36:57 crc kubenswrapper[4733]: I0202 15:36:57.514775 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-eb2d-account-create-update-vzhkk" Feb 02 15:36:58 crc kubenswrapper[4733]: I0202 15:36:58.774611 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["manila-kuttl-tests/manila-db-sync-tdp6r"] Feb 02 15:36:58 crc kubenswrapper[4733]: E0202 15:36:58.774980 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d57375e-7e2f-4870-8c60-555921f94b5e" containerName="manila-scheduler" Feb 02 15:36:58 crc kubenswrapper[4733]: I0202 15:36:58.775000 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d57375e-7e2f-4870-8c60-555921f94b5e" containerName="manila-scheduler" Feb 02 15:36:58 crc kubenswrapper[4733]: E0202 15:36:58.775021 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d04d46e8-50b1-47ab-ac62-a7f446949411" containerName="manila-scheduler" Feb 02 15:36:58 crc kubenswrapper[4733]: I0202 15:36:58.775031 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="d04d46e8-50b1-47ab-ac62-a7f446949411" containerName="manila-scheduler" Feb 02 15:36:58 crc kubenswrapper[4733]: E0202 15:36:58.775046 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78bb94e9-3aad-4917-8a42-105b948191a6" containerName="probe" Feb 02 15:36:58 crc kubenswrapper[4733]: I0202 15:36:58.775057 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="78bb94e9-3aad-4917-8a42-105b948191a6" containerName="probe" Feb 02 15:36:58 crc kubenswrapper[4733]: E0202 15:36:58.775069 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d04d46e8-50b1-47ab-ac62-a7f446949411" containerName="probe" Feb 02 15:36:58 crc kubenswrapper[4733]: I0202 15:36:58.775078 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="d04d46e8-50b1-47ab-ac62-a7f446949411" containerName="probe" Feb 02 15:36:58 crc kubenswrapper[4733]: E0202 15:36:58.775094 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2761761e-fb0e-43f8-8ad9-5835b5f7e619" containerName="mariadb-database-create" Feb 02 15:36:58 crc kubenswrapper[4733]: I0202 15:36:58.775104 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="2761761e-fb0e-43f8-8ad9-5835b5f7e619" containerName="mariadb-database-create" Feb 02 15:36:58 crc kubenswrapper[4733]: E0202 15:36:58.775124 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78bb94e9-3aad-4917-8a42-105b948191a6" containerName="manila-scheduler" Feb 02 15:36:58 crc kubenswrapper[4733]: I0202 15:36:58.775133 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="78bb94e9-3aad-4917-8a42-105b948191a6" containerName="manila-scheduler" Feb 02 15:36:58 crc kubenswrapper[4733]: E0202 15:36:58.775146 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d57375e-7e2f-4870-8c60-555921f94b5e" containerName="probe" Feb 02 15:36:58 crc kubenswrapper[4733]: I0202 15:36:58.775183 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d57375e-7e2f-4870-8c60-555921f94b5e" containerName="probe" Feb 02 15:36:58 crc kubenswrapper[4733]: E0202 15:36:58.775196 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="594f5269-2f23-4b68-8339-3871e122a649" containerName="mariadb-account-create-update" Feb 02 15:36:58 crc kubenswrapper[4733]: I0202 15:36:58.775206 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="594f5269-2f23-4b68-8339-3871e122a649" containerName="mariadb-account-create-update" Feb 02 15:36:58 crc kubenswrapper[4733]: I0202 15:36:58.775345 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="594f5269-2f23-4b68-8339-3871e122a649" containerName="mariadb-account-create-update" Feb 02 15:36:58 crc kubenswrapper[4733]: I0202 15:36:58.775364 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="d04d46e8-50b1-47ab-ac62-a7f446949411" containerName="probe" Feb 02 15:36:58 crc kubenswrapper[4733]: I0202 15:36:58.775374 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="78bb94e9-3aad-4917-8a42-105b948191a6" containerName="probe" Feb 02 15:36:58 crc kubenswrapper[4733]: I0202 15:36:58.775384 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="2761761e-fb0e-43f8-8ad9-5835b5f7e619" containerName="mariadb-database-create" Feb 02 15:36:58 crc kubenswrapper[4733]: I0202 15:36:58.775397 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d57375e-7e2f-4870-8c60-555921f94b5e" containerName="probe" Feb 02 15:36:58 crc kubenswrapper[4733]: I0202 15:36:58.775408 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="d04d46e8-50b1-47ab-ac62-a7f446949411" containerName="manila-scheduler" Feb 02 15:36:58 crc kubenswrapper[4733]: I0202 15:36:58.775423 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d57375e-7e2f-4870-8c60-555921f94b5e" containerName="manila-scheduler" Feb 02 15:36:58 crc kubenswrapper[4733]: I0202 15:36:58.775431 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="78bb94e9-3aad-4917-8a42-105b948191a6" containerName="manila-scheduler" Feb 02 15:36:58 crc kubenswrapper[4733]: I0202 15:36:58.775908 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-db-sync-tdp6r" Feb 02 15:36:58 crc kubenswrapper[4733]: I0202 15:36:58.778216 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"combined-ca-bundle" Feb 02 15:36:58 crc kubenswrapper[4733]: I0202 15:36:58.778507 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"manila-config-data" Feb 02 15:36:58 crc kubenswrapper[4733]: I0202 15:36:58.778585 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"manila-manila-dockercfg-qdrwn" Feb 02 15:36:58 crc kubenswrapper[4733]: I0202 15:36:58.796591 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/manila-db-sync-tdp6r"] Feb 02 15:36:58 crc kubenswrapper[4733]: I0202 15:36:58.941408 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7vvr\" (UniqueName: \"kubernetes.io/projected/8eabd41d-3710-4084-8dc9-40dc4c8f8e97-kube-api-access-r7vvr\") pod \"manila-db-sync-tdp6r\" (UID: \"8eabd41d-3710-4084-8dc9-40dc4c8f8e97\") " pod="manila-kuttl-tests/manila-db-sync-tdp6r" Feb 02 15:36:58 crc kubenswrapper[4733]: I0202 15:36:58.941471 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8eabd41d-3710-4084-8dc9-40dc4c8f8e97-combined-ca-bundle\") pod \"manila-db-sync-tdp6r\" (UID: \"8eabd41d-3710-4084-8dc9-40dc4c8f8e97\") " pod="manila-kuttl-tests/manila-db-sync-tdp6r" Feb 02 15:36:58 crc kubenswrapper[4733]: I0202 15:36:58.941496 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8eabd41d-3710-4084-8dc9-40dc4c8f8e97-config-data\") pod \"manila-db-sync-tdp6r\" (UID: \"8eabd41d-3710-4084-8dc9-40dc4c8f8e97\") " pod="manila-kuttl-tests/manila-db-sync-tdp6r" Feb 02 15:36:58 crc kubenswrapper[4733]: I0202 15:36:58.941513 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/8eabd41d-3710-4084-8dc9-40dc4c8f8e97-job-config-data\") pod \"manila-db-sync-tdp6r\" (UID: \"8eabd41d-3710-4084-8dc9-40dc4c8f8e97\") " pod="manila-kuttl-tests/manila-db-sync-tdp6r" Feb 02 15:36:59 crc kubenswrapper[4733]: I0202 15:36:59.043537 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8eabd41d-3710-4084-8dc9-40dc4c8f8e97-combined-ca-bundle\") pod \"manila-db-sync-tdp6r\" (UID: \"8eabd41d-3710-4084-8dc9-40dc4c8f8e97\") " pod="manila-kuttl-tests/manila-db-sync-tdp6r" Feb 02 15:36:59 crc kubenswrapper[4733]: I0202 15:36:59.043951 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8eabd41d-3710-4084-8dc9-40dc4c8f8e97-config-data\") pod \"manila-db-sync-tdp6r\" (UID: \"8eabd41d-3710-4084-8dc9-40dc4c8f8e97\") " pod="manila-kuttl-tests/manila-db-sync-tdp6r" Feb 02 15:36:59 crc kubenswrapper[4733]: I0202 15:36:59.044109 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/8eabd41d-3710-4084-8dc9-40dc4c8f8e97-job-config-data\") pod \"manila-db-sync-tdp6r\" (UID: \"8eabd41d-3710-4084-8dc9-40dc4c8f8e97\") " pod="manila-kuttl-tests/manila-db-sync-tdp6r" Feb 02 15:36:59 crc kubenswrapper[4733]: I0202 15:36:59.044508 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7vvr\" (UniqueName: \"kubernetes.io/projected/8eabd41d-3710-4084-8dc9-40dc4c8f8e97-kube-api-access-r7vvr\") pod \"manila-db-sync-tdp6r\" (UID: \"8eabd41d-3710-4084-8dc9-40dc4c8f8e97\") " pod="manila-kuttl-tests/manila-db-sync-tdp6r" Feb 02 15:36:59 crc kubenswrapper[4733]: I0202 15:36:59.051345 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/8eabd41d-3710-4084-8dc9-40dc4c8f8e97-job-config-data\") pod \"manila-db-sync-tdp6r\" (UID: \"8eabd41d-3710-4084-8dc9-40dc4c8f8e97\") " pod="manila-kuttl-tests/manila-db-sync-tdp6r" Feb 02 15:36:59 crc kubenswrapper[4733]: I0202 15:36:59.051390 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8eabd41d-3710-4084-8dc9-40dc4c8f8e97-config-data\") pod \"manila-db-sync-tdp6r\" (UID: \"8eabd41d-3710-4084-8dc9-40dc4c8f8e97\") " pod="manila-kuttl-tests/manila-db-sync-tdp6r" Feb 02 15:36:59 crc kubenswrapper[4733]: I0202 15:36:59.051833 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8eabd41d-3710-4084-8dc9-40dc4c8f8e97-combined-ca-bundle\") pod \"manila-db-sync-tdp6r\" (UID: \"8eabd41d-3710-4084-8dc9-40dc4c8f8e97\") " pod="manila-kuttl-tests/manila-db-sync-tdp6r" Feb 02 15:36:59 crc kubenswrapper[4733]: I0202 15:36:59.077710 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7vvr\" (UniqueName: \"kubernetes.io/projected/8eabd41d-3710-4084-8dc9-40dc4c8f8e97-kube-api-access-r7vvr\") pod \"manila-db-sync-tdp6r\" (UID: \"8eabd41d-3710-4084-8dc9-40dc4c8f8e97\") " pod="manila-kuttl-tests/manila-db-sync-tdp6r" Feb 02 15:36:59 crc kubenswrapper[4733]: I0202 15:36:59.096764 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-db-sync-tdp6r" Feb 02 15:36:59 crc kubenswrapper[4733]: I0202 15:36:59.528592 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/manila-db-sync-tdp6r"] Feb 02 15:36:59 crc kubenswrapper[4733]: W0202 15:36:59.533472 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8eabd41d_3710_4084_8dc9_40dc4c8f8e97.slice/crio-7a2f1caee2691ea08e3d3ec81944eb1f7ed2c5586c7b74134f11788290b5b3a9 WatchSource:0}: Error finding container 7a2f1caee2691ea08e3d3ec81944eb1f7ed2c5586c7b74134f11788290b5b3a9: Status 404 returned error can't find the container with id 7a2f1caee2691ea08e3d3ec81944eb1f7ed2c5586c7b74134f11788290b5b3a9 Feb 02 15:37:00 crc kubenswrapper[4733]: I0202 15:37:00.534694 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-db-sync-tdp6r" event={"ID":"8eabd41d-3710-4084-8dc9-40dc4c8f8e97","Type":"ContainerStarted","Data":"1d64e3457cc83f08629d107d40db515c344d361ad519db210f6a4ea6c37f5b37"} Feb 02 15:37:00 crc kubenswrapper[4733]: I0202 15:37:00.535113 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-db-sync-tdp6r" event={"ID":"8eabd41d-3710-4084-8dc9-40dc4c8f8e97","Type":"ContainerStarted","Data":"7a2f1caee2691ea08e3d3ec81944eb1f7ed2c5586c7b74134f11788290b5b3a9"} Feb 02 15:37:00 crc kubenswrapper[4733]: I0202 15:37:00.564091 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="manila-kuttl-tests/manila-db-sync-tdp6r" podStartSLOduration=2.564054779 podStartE2EDuration="2.564054779s" podCreationTimestamp="2026-02-02 15:36:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:37:00.561326328 +0000 UTC m=+1304.012787686" watchObservedRunningTime="2026-02-02 15:37:00.564054779 +0000 UTC m=+1304.015516167" Feb 02 15:37:01 crc kubenswrapper[4733]: I0202 15:37:01.542662 4733 generic.go:334] "Generic (PLEG): container finished" podID="8eabd41d-3710-4084-8dc9-40dc4c8f8e97" containerID="1d64e3457cc83f08629d107d40db515c344d361ad519db210f6a4ea6c37f5b37" exitCode=0 Feb 02 15:37:01 crc kubenswrapper[4733]: I0202 15:37:01.542770 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-db-sync-tdp6r" event={"ID":"8eabd41d-3710-4084-8dc9-40dc4c8f8e97","Type":"ContainerDied","Data":"1d64e3457cc83f08629d107d40db515c344d361ad519db210f6a4ea6c37f5b37"} Feb 02 15:37:02 crc kubenswrapper[4733]: I0202 15:37:02.815126 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-db-sync-tdp6r" Feb 02 15:37:02 crc kubenswrapper[4733]: I0202 15:37:02.899952 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8eabd41d-3710-4084-8dc9-40dc4c8f8e97-config-data\") pod \"8eabd41d-3710-4084-8dc9-40dc4c8f8e97\" (UID: \"8eabd41d-3710-4084-8dc9-40dc4c8f8e97\") " Feb 02 15:37:02 crc kubenswrapper[4733]: I0202 15:37:02.900192 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8eabd41d-3710-4084-8dc9-40dc4c8f8e97-combined-ca-bundle\") pod \"8eabd41d-3710-4084-8dc9-40dc4c8f8e97\" (UID: \"8eabd41d-3710-4084-8dc9-40dc4c8f8e97\") " Feb 02 15:37:02 crc kubenswrapper[4733]: I0202 15:37:02.900227 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7vvr\" (UniqueName: \"kubernetes.io/projected/8eabd41d-3710-4084-8dc9-40dc4c8f8e97-kube-api-access-r7vvr\") pod \"8eabd41d-3710-4084-8dc9-40dc4c8f8e97\" (UID: \"8eabd41d-3710-4084-8dc9-40dc4c8f8e97\") " Feb 02 15:37:02 crc kubenswrapper[4733]: I0202 15:37:02.900250 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/8eabd41d-3710-4084-8dc9-40dc4c8f8e97-job-config-data\") pod \"8eabd41d-3710-4084-8dc9-40dc4c8f8e97\" (UID: \"8eabd41d-3710-4084-8dc9-40dc4c8f8e97\") " Feb 02 15:37:02 crc kubenswrapper[4733]: I0202 15:37:02.906001 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8eabd41d-3710-4084-8dc9-40dc4c8f8e97-kube-api-access-r7vvr" (OuterVolumeSpecName: "kube-api-access-r7vvr") pod "8eabd41d-3710-4084-8dc9-40dc4c8f8e97" (UID: "8eabd41d-3710-4084-8dc9-40dc4c8f8e97"). InnerVolumeSpecName "kube-api-access-r7vvr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:37:02 crc kubenswrapper[4733]: I0202 15:37:02.906094 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8eabd41d-3710-4084-8dc9-40dc4c8f8e97-job-config-data" (OuterVolumeSpecName: "job-config-data") pod "8eabd41d-3710-4084-8dc9-40dc4c8f8e97" (UID: "8eabd41d-3710-4084-8dc9-40dc4c8f8e97"). InnerVolumeSpecName "job-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:37:02 crc kubenswrapper[4733]: I0202 15:37:02.909103 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8eabd41d-3710-4084-8dc9-40dc4c8f8e97-config-data" (OuterVolumeSpecName: "config-data") pod "8eabd41d-3710-4084-8dc9-40dc4c8f8e97" (UID: "8eabd41d-3710-4084-8dc9-40dc4c8f8e97"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:37:02 crc kubenswrapper[4733]: I0202 15:37:02.921919 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8eabd41d-3710-4084-8dc9-40dc4c8f8e97-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8eabd41d-3710-4084-8dc9-40dc4c8f8e97" (UID: "8eabd41d-3710-4084-8dc9-40dc4c8f8e97"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:37:03 crc kubenswrapper[4733]: I0202 15:37:03.001841 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8eabd41d-3710-4084-8dc9-40dc4c8f8e97-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 15:37:03 crc kubenswrapper[4733]: I0202 15:37:03.001875 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7vvr\" (UniqueName: \"kubernetes.io/projected/8eabd41d-3710-4084-8dc9-40dc4c8f8e97-kube-api-access-r7vvr\") on node \"crc\" DevicePath \"\"" Feb 02 15:37:03 crc kubenswrapper[4733]: I0202 15:37:03.001887 4733 reconciler_common.go:293] "Volume detached for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/8eabd41d-3710-4084-8dc9-40dc4c8f8e97-job-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 15:37:03 crc kubenswrapper[4733]: I0202 15:37:03.001897 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8eabd41d-3710-4084-8dc9-40dc4c8f8e97-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 15:37:03 crc kubenswrapper[4733]: I0202 15:37:03.574827 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-db-sync-tdp6r" event={"ID":"8eabd41d-3710-4084-8dc9-40dc4c8f8e97","Type":"ContainerDied","Data":"7a2f1caee2691ea08e3d3ec81944eb1f7ed2c5586c7b74134f11788290b5b3a9"} Feb 02 15:37:03 crc kubenswrapper[4733]: I0202 15:37:03.574864 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a2f1caee2691ea08e3d3ec81944eb1f7ed2c5586c7b74134f11788290b5b3a9" Feb 02 15:37:03 crc kubenswrapper[4733]: I0202 15:37:03.574909 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-db-sync-tdp6r" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.093107 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["manila-kuttl-tests/manila-scheduler-0"] Feb 02 15:37:04 crc kubenswrapper[4733]: E0202 15:37:04.093799 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8eabd41d-3710-4084-8dc9-40dc4c8f8e97" containerName="manila-db-sync" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.093817 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="8eabd41d-3710-4084-8dc9-40dc4c8f8e97" containerName="manila-db-sync" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.093989 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="8eabd41d-3710-4084-8dc9-40dc4c8f8e97" containerName="manila-db-sync" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.094815 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.101198 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"combined-ca-bundle" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.101710 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"manila-config-data" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.101964 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"manila-scheduler-config-data" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.102352 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"manila-manila-dockercfg-qdrwn" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.102602 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"manila-scripts" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.113007 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/manila-scheduler-0"] Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.140434 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["manila-kuttl-tests/manila-share-share0-0"] Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.141547 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.144518 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"manila-share-share0-config-data" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.144750 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"ceph-conf-files" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.146374 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/manila-share-share0-0"] Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.278723 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["manila-kuttl-tests/manila-api-0"] Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.279644 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-api-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.284028 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"cert-manila-internal-svc" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.284266 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"cert-manila-public-svc" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.284411 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"manila-api-config-data" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.285317 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bf8aaefd-448c-4b30-beb7-6590b35a5209-etc-machine-id\") pod \"manila-share-share0-0\" (UID: \"bf8aaefd-448c-4b30-beb7-6590b35a5209\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.285348 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bf8aaefd-448c-4b30-beb7-6590b35a5209-config-data-custom\") pod \"manila-share-share0-0\" (UID: \"bf8aaefd-448c-4b30-beb7-6590b35a5209\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.285377 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bf8aaefd-448c-4b30-beb7-6590b35a5209-ceph\") pod \"manila-share-share0-0\" (UID: \"bf8aaefd-448c-4b30-beb7-6590b35a5209\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.285405 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/bf8aaefd-448c-4b30-beb7-6590b35a5209-var-lib-manila\") pod \"manila-share-share0-0\" (UID: \"bf8aaefd-448c-4b30-beb7-6590b35a5209\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.285424 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eeaf18e2-5cfc-4795-9f47-5c42f217c7cb-scripts\") pod \"manila-scheduler-0\" (UID: \"eeaf18e2-5cfc-4795-9f47-5c42f217c7cb\") " pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.285450 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dk5gw\" (UniqueName: \"kubernetes.io/projected/eeaf18e2-5cfc-4795-9f47-5c42f217c7cb-kube-api-access-dk5gw\") pod \"manila-scheduler-0\" (UID: \"eeaf18e2-5cfc-4795-9f47-5c42f217c7cb\") " pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.285480 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eeaf18e2-5cfc-4795-9f47-5c42f217c7cb-config-data\") pod \"manila-scheduler-0\" (UID: \"eeaf18e2-5cfc-4795-9f47-5c42f217c7cb\") " pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.285508 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf8aaefd-448c-4b30-beb7-6590b35a5209-scripts\") pod \"manila-share-share0-0\" (UID: \"bf8aaefd-448c-4b30-beb7-6590b35a5209\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.285527 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eeaf18e2-5cfc-4795-9f47-5c42f217c7cb-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"eeaf18e2-5cfc-4795-9f47-5c42f217c7cb\") " pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.285541 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf8aaefd-448c-4b30-beb7-6590b35a5209-config-data\") pod \"manila-share-share0-0\" (UID: \"bf8aaefd-448c-4b30-beb7-6590b35a5209\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.285556 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmqv9\" (UniqueName: \"kubernetes.io/projected/bf8aaefd-448c-4b30-beb7-6590b35a5209-kube-api-access-zmqv9\") pod \"manila-share-share0-0\" (UID: \"bf8aaefd-448c-4b30-beb7-6590b35a5209\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.285573 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eeaf18e2-5cfc-4795-9f47-5c42f217c7cb-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"eeaf18e2-5cfc-4795-9f47-5c42f217c7cb\") " pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.285589 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eeaf18e2-5cfc-4795-9f47-5c42f217c7cb-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"eeaf18e2-5cfc-4795-9f47-5c42f217c7cb\") " pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.285603 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf8aaefd-448c-4b30-beb7-6590b35a5209-combined-ca-bundle\") pod \"manila-share-share0-0\" (UID: \"bf8aaefd-448c-4b30-beb7-6590b35a5209\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.292306 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/manila-api-0"] Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.406600 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4-public-tls-certs\") pod \"manila-api-0\" (UID: \"f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.406679 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/bf8aaefd-448c-4b30-beb7-6590b35a5209-var-lib-manila\") pod \"manila-share-share0-0\" (UID: \"bf8aaefd-448c-4b30-beb7-6590b35a5209\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.406708 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eeaf18e2-5cfc-4795-9f47-5c42f217c7cb-scripts\") pod \"manila-scheduler-0\" (UID: \"eeaf18e2-5cfc-4795-9f47-5c42f217c7cb\") " pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.407651 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dk5gw\" (UniqueName: \"kubernetes.io/projected/eeaf18e2-5cfc-4795-9f47-5c42f217c7cb-kube-api-access-dk5gw\") pod \"manila-scheduler-0\" (UID: \"eeaf18e2-5cfc-4795-9f47-5c42f217c7cb\") " pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.408032 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4-etc-machine-id\") pod \"manila-api-0\" (UID: \"f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.408129 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eeaf18e2-5cfc-4795-9f47-5c42f217c7cb-config-data\") pod \"manila-scheduler-0\" (UID: \"eeaf18e2-5cfc-4795-9f47-5c42f217c7cb\") " pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.408152 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4-scripts\") pod \"manila-api-0\" (UID: \"f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.408310 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4-config-data\") pod \"manila-api-0\" (UID: \"f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.408387 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4-internal-tls-certs\") pod \"manila-api-0\" (UID: \"f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.408409 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4-logs\") pod \"manila-api-0\" (UID: \"f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.408435 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf8aaefd-448c-4b30-beb7-6590b35a5209-scripts\") pod \"manila-share-share0-0\" (UID: \"bf8aaefd-448c-4b30-beb7-6590b35a5209\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.408476 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eeaf18e2-5cfc-4795-9f47-5c42f217c7cb-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"eeaf18e2-5cfc-4795-9f47-5c42f217c7cb\") " pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.408498 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf8aaefd-448c-4b30-beb7-6590b35a5209-config-data\") pod \"manila-share-share0-0\" (UID: \"bf8aaefd-448c-4b30-beb7-6590b35a5209\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.408518 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmqv9\" (UniqueName: \"kubernetes.io/projected/bf8aaefd-448c-4b30-beb7-6590b35a5209-kube-api-access-zmqv9\") pod \"manila-share-share0-0\" (UID: \"bf8aaefd-448c-4b30-beb7-6590b35a5209\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.408584 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eeaf18e2-5cfc-4795-9f47-5c42f217c7cb-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"eeaf18e2-5cfc-4795-9f47-5c42f217c7cb\") " pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.408613 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eeaf18e2-5cfc-4795-9f47-5c42f217c7cb-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"eeaf18e2-5cfc-4795-9f47-5c42f217c7cb\") " pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.408635 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf8aaefd-448c-4b30-beb7-6590b35a5209-combined-ca-bundle\") pod \"manila-share-share0-0\" (UID: \"bf8aaefd-448c-4b30-beb7-6590b35a5209\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.408669 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4-config-data-custom\") pod \"manila-api-0\" (UID: \"f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.408751 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzm8n\" (UniqueName: \"kubernetes.io/projected/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4-kube-api-access-hzm8n\") pod \"manila-api-0\" (UID: \"f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.408786 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bf8aaefd-448c-4b30-beb7-6590b35a5209-etc-machine-id\") pod \"manila-share-share0-0\" (UID: \"bf8aaefd-448c-4b30-beb7-6590b35a5209\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.408808 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bf8aaefd-448c-4b30-beb7-6590b35a5209-config-data-custom\") pod \"manila-share-share0-0\" (UID: \"bf8aaefd-448c-4b30-beb7-6590b35a5209\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.413480 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eeaf18e2-5cfc-4795-9f47-5c42f217c7cb-scripts\") pod \"manila-scheduler-0\" (UID: \"eeaf18e2-5cfc-4795-9f47-5c42f217c7cb\") " pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.415651 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bf8aaefd-448c-4b30-beb7-6590b35a5209-ceph\") pod \"manila-share-share0-0\" (UID: \"bf8aaefd-448c-4b30-beb7-6590b35a5209\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.415709 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.417569 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bf8aaefd-448c-4b30-beb7-6590b35a5209-etc-machine-id\") pod \"manila-share-share0-0\" (UID: \"bf8aaefd-448c-4b30-beb7-6590b35a5209\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.421747 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eeaf18e2-5cfc-4795-9f47-5c42f217c7cb-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"eeaf18e2-5cfc-4795-9f47-5c42f217c7cb\") " pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.422946 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf8aaefd-448c-4b30-beb7-6590b35a5209-scripts\") pod \"manila-share-share0-0\" (UID: \"bf8aaefd-448c-4b30-beb7-6590b35a5209\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.424310 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bf8aaefd-448c-4b30-beb7-6590b35a5209-ceph\") pod \"manila-share-share0-0\" (UID: \"bf8aaefd-448c-4b30-beb7-6590b35a5209\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.407703 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/bf8aaefd-448c-4b30-beb7-6590b35a5209-var-lib-manila\") pod \"manila-share-share0-0\" (UID: \"bf8aaefd-448c-4b30-beb7-6590b35a5209\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.424974 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eeaf18e2-5cfc-4795-9f47-5c42f217c7cb-config-data\") pod \"manila-scheduler-0\" (UID: \"eeaf18e2-5cfc-4795-9f47-5c42f217c7cb\") " pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.448780 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eeaf18e2-5cfc-4795-9f47-5c42f217c7cb-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"eeaf18e2-5cfc-4795-9f47-5c42f217c7cb\") " pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.450997 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf8aaefd-448c-4b30-beb7-6590b35a5209-config-data\") pod \"manila-share-share0-0\" (UID: \"bf8aaefd-448c-4b30-beb7-6590b35a5209\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.451625 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bf8aaefd-448c-4b30-beb7-6590b35a5209-config-data-custom\") pod \"manila-share-share0-0\" (UID: \"bf8aaefd-448c-4b30-beb7-6590b35a5209\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.452667 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf8aaefd-448c-4b30-beb7-6590b35a5209-combined-ca-bundle\") pod \"manila-share-share0-0\" (UID: \"bf8aaefd-448c-4b30-beb7-6590b35a5209\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.453589 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmqv9\" (UniqueName: \"kubernetes.io/projected/bf8aaefd-448c-4b30-beb7-6590b35a5209-kube-api-access-zmqv9\") pod \"manila-share-share0-0\" (UID: \"bf8aaefd-448c-4b30-beb7-6590b35a5209\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.453727 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eeaf18e2-5cfc-4795-9f47-5c42f217c7cb-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"eeaf18e2-5cfc-4795-9f47-5c42f217c7cb\") " pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.455843 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dk5gw\" (UniqueName: \"kubernetes.io/projected/eeaf18e2-5cfc-4795-9f47-5c42f217c7cb-kube-api-access-dk5gw\") pod \"manila-scheduler-0\" (UID: \"eeaf18e2-5cfc-4795-9f47-5c42f217c7cb\") " pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.456946 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.518880 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzm8n\" (UniqueName: \"kubernetes.io/projected/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4-kube-api-access-hzm8n\") pod \"manila-api-0\" (UID: \"f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.519063 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.519099 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4-public-tls-certs\") pod \"manila-api-0\" (UID: \"f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.519145 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4-etc-machine-id\") pod \"manila-api-0\" (UID: \"f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.519200 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4-scripts\") pod \"manila-api-0\" (UID: \"f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.519227 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4-config-data\") pod \"manila-api-0\" (UID: \"f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.519256 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4-internal-tls-certs\") pod \"manila-api-0\" (UID: \"f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.519278 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4-logs\") pod \"manila-api-0\" (UID: \"f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.519317 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4-config-data-custom\") pod \"manila-api-0\" (UID: \"f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.519905 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4-etc-machine-id\") pod \"manila-api-0\" (UID: \"f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.523612 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4-logs\") pod \"manila-api-0\" (UID: \"f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.523936 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.524336 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4-config-data\") pod \"manila-api-0\" (UID: \"f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.524924 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4-internal-tls-certs\") pod \"manila-api-0\" (UID: \"f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.525076 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4-config-data-custom\") pod \"manila-api-0\" (UID: \"f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.526037 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4-scripts\") pod \"manila-api-0\" (UID: \"f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.529404 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4-public-tls-certs\") pod \"manila-api-0\" (UID: \"f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.540832 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzm8n\" (UniqueName: \"kubernetes.io/projected/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4-kube-api-access-hzm8n\") pod \"manila-api-0\" (UID: \"f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.607083 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-api-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.723876 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.818207 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/manila-api-0"] Feb 02 15:37:04 crc kubenswrapper[4733]: W0202 15:37:04.823287 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf8d44cee_b95d_40e0_9cfc_ebf3f7b87db4.slice/crio-c745e7444ca635bd5b159a1c60f5ec0c435d991afc72b2586220a0998f7b3c41 WatchSource:0}: Error finding container c745e7444ca635bd5b159a1c60f5ec0c435d991afc72b2586220a0998f7b3c41: Status 404 returned error can't find the container with id c745e7444ca635bd5b159a1c60f5ec0c435d991afc72b2586220a0998f7b3c41 Feb 02 15:37:04 crc kubenswrapper[4733]: I0202 15:37:04.910368 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/manila-share-share0-0"] Feb 02 15:37:05 crc kubenswrapper[4733]: I0202 15:37:05.150053 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/manila-scheduler-0"] Feb 02 15:37:05 crc kubenswrapper[4733]: I0202 15:37:05.592462 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-share-share0-0" event={"ID":"bf8aaefd-448c-4b30-beb7-6590b35a5209","Type":"ContainerStarted","Data":"017614a4cc5c9d894b40b309a350ff8d25e8495d0c82dcb2a7ab58c0802f74dc"} Feb 02 15:37:05 crc kubenswrapper[4733]: I0202 15:37:05.593085 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-share-share0-0" event={"ID":"bf8aaefd-448c-4b30-beb7-6590b35a5209","Type":"ContainerStarted","Data":"90bcc706ad9b965694a7c31750b565b29e30aec7d714cbef5034451597ed42e8"} Feb 02 15:37:05 crc kubenswrapper[4733]: I0202 15:37:05.595312 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-scheduler-0" event={"ID":"eeaf18e2-5cfc-4795-9f47-5c42f217c7cb","Type":"ContainerStarted","Data":"a1c73b28f4ba7af7cc6b7f195dee621cd2d41b88829f92a723252238e31b3ab2"} Feb 02 15:37:05 crc kubenswrapper[4733]: I0202 15:37:05.595345 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-scheduler-0" event={"ID":"eeaf18e2-5cfc-4795-9f47-5c42f217c7cb","Type":"ContainerStarted","Data":"187c94bb8a061c783791b8614b7a95fe91a47032fe4be25304c3117dd9430fff"} Feb 02 15:37:05 crc kubenswrapper[4733]: I0202 15:37:05.597751 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-api-0" event={"ID":"f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4","Type":"ContainerStarted","Data":"b98b8e2b776fbd00e5c51fab9b2d2d57009eaf7e21b3d40d65fa1cbbb522f7d2"} Feb 02 15:37:05 crc kubenswrapper[4733]: I0202 15:37:05.597786 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-api-0" event={"ID":"f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4","Type":"ContainerStarted","Data":"c745e7444ca635bd5b159a1c60f5ec0c435d991afc72b2586220a0998f7b3c41"} Feb 02 15:37:06 crc kubenswrapper[4733]: I0202 15:37:06.610556 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-share-share0-0" event={"ID":"bf8aaefd-448c-4b30-beb7-6590b35a5209","Type":"ContainerStarted","Data":"80a73e88fd8eb4d9f137038a5a6508422dcfc4328e5d69431221cc73992df3f4"} Feb 02 15:37:06 crc kubenswrapper[4733]: I0202 15:37:06.612741 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-scheduler-0" event={"ID":"eeaf18e2-5cfc-4795-9f47-5c42f217c7cb","Type":"ContainerStarted","Data":"ea5d03fc685f9d6312eb6363d820b95b9da351a7b850efb702d75bd4714f225a"} Feb 02 15:37:06 crc kubenswrapper[4733]: I0202 15:37:06.614729 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-api-0" event={"ID":"f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4","Type":"ContainerStarted","Data":"8e5d40a60ab730b2ebb52872f09f32bf64a8e1c12a566815bdf9eb1eec77d9cb"} Feb 02 15:37:06 crc kubenswrapper[4733]: I0202 15:37:06.614927 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="manila-kuttl-tests/manila-api-0" Feb 02 15:37:06 crc kubenswrapper[4733]: I0202 15:37:06.643436 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="manila-kuttl-tests/manila-share-share0-0" podStartSLOduration=2.643420553 podStartE2EDuration="2.643420553s" podCreationTimestamp="2026-02-02 15:37:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:37:06.637073558 +0000 UTC m=+1310.088534936" watchObservedRunningTime="2026-02-02 15:37:06.643420553 +0000 UTC m=+1310.094881911" Feb 02 15:37:06 crc kubenswrapper[4733]: I0202 15:37:06.663256 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="manila-kuttl-tests/manila-scheduler-0" podStartSLOduration=2.663234497 podStartE2EDuration="2.663234497s" podCreationTimestamp="2026-02-02 15:37:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:37:06.661320107 +0000 UTC m=+1310.112781465" watchObservedRunningTime="2026-02-02 15:37:06.663234497 +0000 UTC m=+1310.114695865" Feb 02 15:37:06 crc kubenswrapper[4733]: I0202 15:37:06.686397 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="manila-kuttl-tests/manila-api-0" podStartSLOduration=2.686379129 podStartE2EDuration="2.686379129s" podCreationTimestamp="2026-02-02 15:37:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:37:06.683102643 +0000 UTC m=+1310.134564001" watchObservedRunningTime="2026-02-02 15:37:06.686379129 +0000 UTC m=+1310.137840487" Feb 02 15:37:14 crc kubenswrapper[4733]: I0202 15:37:14.458059 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:37:14 crc kubenswrapper[4733]: I0202 15:37:14.724040 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:37:25 crc kubenswrapper[4733]: I0202 15:37:25.874942 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="manila-kuttl-tests/manila-api-0" Feb 02 15:37:26 crc kubenswrapper[4733]: I0202 15:37:26.074008 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:37:26 crc kubenswrapper[4733]: I0202 15:37:26.188918 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:37:26 crc kubenswrapper[4733]: I0202 15:37:26.808018 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/manila-db-sync-tdp6r"] Feb 02 15:37:26 crc kubenswrapper[4733]: I0202 15:37:26.819970 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["manila-kuttl-tests/manila-db-sync-tdp6r"] Feb 02 15:37:26 crc kubenswrapper[4733]: I0202 15:37:26.835071 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/manila-share-share0-0"] Feb 02 15:37:26 crc kubenswrapper[4733]: I0202 15:37:26.835406 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="manila-kuttl-tests/manila-share-share0-0" podUID="bf8aaefd-448c-4b30-beb7-6590b35a5209" containerName="manila-share" containerID="cri-o://017614a4cc5c9d894b40b309a350ff8d25e8495d0c82dcb2a7ab58c0802f74dc" gracePeriod=30 Feb 02 15:37:26 crc kubenswrapper[4733]: I0202 15:37:26.835475 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="manila-kuttl-tests/manila-share-share0-0" podUID="bf8aaefd-448c-4b30-beb7-6590b35a5209" containerName="probe" containerID="cri-o://80a73e88fd8eb4d9f137038a5a6508422dcfc4328e5d69431221cc73992df3f4" gracePeriod=30 Feb 02 15:37:26 crc kubenswrapper[4733]: I0202 15:37:26.848033 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/manila-scheduler-0"] Feb 02 15:37:26 crc kubenswrapper[4733]: I0202 15:37:26.848677 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="manila-kuttl-tests/manila-scheduler-0" podUID="eeaf18e2-5cfc-4795-9f47-5c42f217c7cb" containerName="manila-scheduler" containerID="cri-o://a1c73b28f4ba7af7cc6b7f195dee621cd2d41b88829f92a723252238e31b3ab2" gracePeriod=30 Feb 02 15:37:26 crc kubenswrapper[4733]: I0202 15:37:26.849043 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="manila-kuttl-tests/manila-scheduler-0" podUID="eeaf18e2-5cfc-4795-9f47-5c42f217c7cb" containerName="probe" containerID="cri-o://ea5d03fc685f9d6312eb6363d820b95b9da351a7b850efb702d75bd4714f225a" gracePeriod=30 Feb 02 15:37:26 crc kubenswrapper[4733]: I0202 15:37:26.868462 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["manila-kuttl-tests/manilaeb2d-account-delete-kx6cd"] Feb 02 15:37:26 crc kubenswrapper[4733]: I0202 15:37:26.869462 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manilaeb2d-account-delete-kx6cd" Feb 02 15:37:26 crc kubenswrapper[4733]: I0202 15:37:26.891832 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/manilaeb2d-account-delete-kx6cd"] Feb 02 15:37:26 crc kubenswrapper[4733]: I0202 15:37:26.926951 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/manila-api-0"] Feb 02 15:37:26 crc kubenswrapper[4733]: I0202 15:37:26.927230 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="manila-kuttl-tests/manila-api-0" podUID="f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4" containerName="manila-api-log" containerID="cri-o://b98b8e2b776fbd00e5c51fab9b2d2d57009eaf7e21b3d40d65fa1cbbb522f7d2" gracePeriod=30 Feb 02 15:37:26 crc kubenswrapper[4733]: I0202 15:37:26.927305 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="manila-kuttl-tests/manila-api-0" podUID="f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4" containerName="manila-api" containerID="cri-o://8e5d40a60ab730b2ebb52872f09f32bf64a8e1c12a566815bdf9eb1eec77d9cb" gracePeriod=30 Feb 02 15:37:26 crc kubenswrapper[4733]: I0202 15:37:26.969980 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/27a3df30-36df-4380-b693-18775d6b573a-operator-scripts\") pod \"manilaeb2d-account-delete-kx6cd\" (UID: \"27a3df30-36df-4380-b693-18775d6b573a\") " pod="manila-kuttl-tests/manilaeb2d-account-delete-kx6cd" Feb 02 15:37:26 crc kubenswrapper[4733]: I0202 15:37:26.970041 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddcws\" (UniqueName: \"kubernetes.io/projected/27a3df30-36df-4380-b693-18775d6b573a-kube-api-access-ddcws\") pod \"manilaeb2d-account-delete-kx6cd\" (UID: \"27a3df30-36df-4380-b693-18775d6b573a\") " pod="manila-kuttl-tests/manilaeb2d-account-delete-kx6cd" Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.071993 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/27a3df30-36df-4380-b693-18775d6b573a-operator-scripts\") pod \"manilaeb2d-account-delete-kx6cd\" (UID: \"27a3df30-36df-4380-b693-18775d6b573a\") " pod="manila-kuttl-tests/manilaeb2d-account-delete-kx6cd" Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.072047 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddcws\" (UniqueName: \"kubernetes.io/projected/27a3df30-36df-4380-b693-18775d6b573a-kube-api-access-ddcws\") pod \"manilaeb2d-account-delete-kx6cd\" (UID: \"27a3df30-36df-4380-b693-18775d6b573a\") " pod="manila-kuttl-tests/manilaeb2d-account-delete-kx6cd" Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.073149 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/27a3df30-36df-4380-b693-18775d6b573a-operator-scripts\") pod \"manilaeb2d-account-delete-kx6cd\" (UID: \"27a3df30-36df-4380-b693-18775d6b573a\") " pod="manila-kuttl-tests/manilaeb2d-account-delete-kx6cd" Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.092868 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddcws\" (UniqueName: \"kubernetes.io/projected/27a3df30-36df-4380-b693-18775d6b573a-kube-api-access-ddcws\") pod \"manilaeb2d-account-delete-kx6cd\" (UID: \"27a3df30-36df-4380-b693-18775d6b573a\") " pod="manila-kuttl-tests/manilaeb2d-account-delete-kx6cd" Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.209508 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manilaeb2d-account-delete-kx6cd" Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.265148 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8eabd41d-3710-4084-8dc9-40dc4c8f8e97" path="/var/lib/kubelet/pods/8eabd41d-3710-4084-8dc9-40dc4c8f8e97/volumes" Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.503548 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/manilaeb2d-account-delete-kx6cd"] Feb 02 15:37:27 crc kubenswrapper[4733]: W0202 15:37:27.514762 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod27a3df30_36df_4380_b693_18775d6b573a.slice/crio-981688b574f254944ebc024b33116050042703ca3a5f6fb05f5a26ad6ed270b5 WatchSource:0}: Error finding container 981688b574f254944ebc024b33116050042703ca3a5f6fb05f5a26ad6ed270b5: Status 404 returned error can't find the container with id 981688b574f254944ebc024b33116050042703ca3a5f6fb05f5a26ad6ed270b5 Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.727439 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.804645 4733 generic.go:334] "Generic (PLEG): container finished" podID="bf8aaefd-448c-4b30-beb7-6590b35a5209" containerID="80a73e88fd8eb4d9f137038a5a6508422dcfc4328e5d69431221cc73992df3f4" exitCode=0 Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.804679 4733 generic.go:334] "Generic (PLEG): container finished" podID="bf8aaefd-448c-4b30-beb7-6590b35a5209" containerID="017614a4cc5c9d894b40b309a350ff8d25e8495d0c82dcb2a7ab58c0802f74dc" exitCode=1 Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.804722 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.804735 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-share-share0-0" event={"ID":"bf8aaefd-448c-4b30-beb7-6590b35a5209","Type":"ContainerDied","Data":"80a73e88fd8eb4d9f137038a5a6508422dcfc4328e5d69431221cc73992df3f4"} Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.804766 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-share-share0-0" event={"ID":"bf8aaefd-448c-4b30-beb7-6590b35a5209","Type":"ContainerDied","Data":"017614a4cc5c9d894b40b309a350ff8d25e8495d0c82dcb2a7ab58c0802f74dc"} Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.804782 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-share-share0-0" event={"ID":"bf8aaefd-448c-4b30-beb7-6590b35a5209","Type":"ContainerDied","Data":"90bcc706ad9b965694a7c31750b565b29e30aec7d714cbef5034451597ed42e8"} Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.804801 4733 scope.go:117] "RemoveContainer" containerID="80a73e88fd8eb4d9f137038a5a6508422dcfc4328e5d69431221cc73992df3f4" Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.808285 4733 generic.go:334] "Generic (PLEG): container finished" podID="eeaf18e2-5cfc-4795-9f47-5c42f217c7cb" containerID="ea5d03fc685f9d6312eb6363d820b95b9da351a7b850efb702d75bd4714f225a" exitCode=0 Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.808342 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-scheduler-0" event={"ID":"eeaf18e2-5cfc-4795-9f47-5c42f217c7cb","Type":"ContainerDied","Data":"ea5d03fc685f9d6312eb6363d820b95b9da351a7b850efb702d75bd4714f225a"} Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.809979 4733 generic.go:334] "Generic (PLEG): container finished" podID="f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4" containerID="b98b8e2b776fbd00e5c51fab9b2d2d57009eaf7e21b3d40d65fa1cbbb522f7d2" exitCode=143 Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.810028 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-api-0" event={"ID":"f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4","Type":"ContainerDied","Data":"b98b8e2b776fbd00e5c51fab9b2d2d57009eaf7e21b3d40d65fa1cbbb522f7d2"} Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.811329 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manilaeb2d-account-delete-kx6cd" event={"ID":"27a3df30-36df-4380-b693-18775d6b573a","Type":"ContainerStarted","Data":"7909cf6a610bf1ebc516cb20c801752e47646be453177b9151b29750fcba7c66"} Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.811353 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manilaeb2d-account-delete-kx6cd" event={"ID":"27a3df30-36df-4380-b693-18775d6b573a","Type":"ContainerStarted","Data":"981688b574f254944ebc024b33116050042703ca3a5f6fb05f5a26ad6ed270b5"} Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.820791 4733 scope.go:117] "RemoveContainer" containerID="017614a4cc5c9d894b40b309a350ff8d25e8495d0c82dcb2a7ab58c0802f74dc" Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.829883 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="manila-kuttl-tests/manilaeb2d-account-delete-kx6cd" podStartSLOduration=1.8298660679999998 podStartE2EDuration="1.829866068s" podCreationTimestamp="2026-02-02 15:37:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:37:27.827652621 +0000 UTC m=+1331.279113979" watchObservedRunningTime="2026-02-02 15:37:27.829866068 +0000 UTC m=+1331.281327426" Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.881904 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bf8aaefd-448c-4b30-beb7-6590b35a5209-config-data-custom\") pod \"bf8aaefd-448c-4b30-beb7-6590b35a5209\" (UID: \"bf8aaefd-448c-4b30-beb7-6590b35a5209\") " Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.881956 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/bf8aaefd-448c-4b30-beb7-6590b35a5209-var-lib-manila\") pod \"bf8aaefd-448c-4b30-beb7-6590b35a5209\" (UID: \"bf8aaefd-448c-4b30-beb7-6590b35a5209\") " Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.881980 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf8aaefd-448c-4b30-beb7-6590b35a5209-config-data\") pod \"bf8aaefd-448c-4b30-beb7-6590b35a5209\" (UID: \"bf8aaefd-448c-4b30-beb7-6590b35a5209\") " Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.882043 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf8aaefd-448c-4b30-beb7-6590b35a5209-scripts\") pod \"bf8aaefd-448c-4b30-beb7-6590b35a5209\" (UID: \"bf8aaefd-448c-4b30-beb7-6590b35a5209\") " Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.882062 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf8aaefd-448c-4b30-beb7-6590b35a5209-combined-ca-bundle\") pod \"bf8aaefd-448c-4b30-beb7-6590b35a5209\" (UID: \"bf8aaefd-448c-4b30-beb7-6590b35a5209\") " Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.882083 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bf8aaefd-448c-4b30-beb7-6590b35a5209-ceph\") pod \"bf8aaefd-448c-4b30-beb7-6590b35a5209\" (UID: \"bf8aaefd-448c-4b30-beb7-6590b35a5209\") " Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.882114 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bf8aaefd-448c-4b30-beb7-6590b35a5209-etc-machine-id\") pod \"bf8aaefd-448c-4b30-beb7-6590b35a5209\" (UID: \"bf8aaefd-448c-4b30-beb7-6590b35a5209\") " Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.882140 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zmqv9\" (UniqueName: \"kubernetes.io/projected/bf8aaefd-448c-4b30-beb7-6590b35a5209-kube-api-access-zmqv9\") pod \"bf8aaefd-448c-4b30-beb7-6590b35a5209\" (UID: \"bf8aaefd-448c-4b30-beb7-6590b35a5209\") " Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.883543 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bf8aaefd-448c-4b30-beb7-6590b35a5209-var-lib-manila" (OuterVolumeSpecName: "var-lib-manila") pod "bf8aaefd-448c-4b30-beb7-6590b35a5209" (UID: "bf8aaefd-448c-4b30-beb7-6590b35a5209"). InnerVolumeSpecName "var-lib-manila". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.883709 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bf8aaefd-448c-4b30-beb7-6590b35a5209-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "bf8aaefd-448c-4b30-beb7-6590b35a5209" (UID: "bf8aaefd-448c-4b30-beb7-6590b35a5209"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.887439 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf8aaefd-448c-4b30-beb7-6590b35a5209-kube-api-access-zmqv9" (OuterVolumeSpecName: "kube-api-access-zmqv9") pod "bf8aaefd-448c-4b30-beb7-6590b35a5209" (UID: "bf8aaefd-448c-4b30-beb7-6590b35a5209"). InnerVolumeSpecName "kube-api-access-zmqv9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.887598 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf8aaefd-448c-4b30-beb7-6590b35a5209-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "bf8aaefd-448c-4b30-beb7-6590b35a5209" (UID: "bf8aaefd-448c-4b30-beb7-6590b35a5209"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.887680 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf8aaefd-448c-4b30-beb7-6590b35a5209-scripts" (OuterVolumeSpecName: "scripts") pod "bf8aaefd-448c-4b30-beb7-6590b35a5209" (UID: "bf8aaefd-448c-4b30-beb7-6590b35a5209"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.889346 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf8aaefd-448c-4b30-beb7-6590b35a5209-ceph" (OuterVolumeSpecName: "ceph") pod "bf8aaefd-448c-4b30-beb7-6590b35a5209" (UID: "bf8aaefd-448c-4b30-beb7-6590b35a5209"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.903205 4733 scope.go:117] "RemoveContainer" containerID="80a73e88fd8eb4d9f137038a5a6508422dcfc4328e5d69431221cc73992df3f4" Feb 02 15:37:27 crc kubenswrapper[4733]: E0202 15:37:27.903711 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80a73e88fd8eb4d9f137038a5a6508422dcfc4328e5d69431221cc73992df3f4\": container with ID starting with 80a73e88fd8eb4d9f137038a5a6508422dcfc4328e5d69431221cc73992df3f4 not found: ID does not exist" containerID="80a73e88fd8eb4d9f137038a5a6508422dcfc4328e5d69431221cc73992df3f4" Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.903741 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80a73e88fd8eb4d9f137038a5a6508422dcfc4328e5d69431221cc73992df3f4"} err="failed to get container status \"80a73e88fd8eb4d9f137038a5a6508422dcfc4328e5d69431221cc73992df3f4\": rpc error: code = NotFound desc = could not find container \"80a73e88fd8eb4d9f137038a5a6508422dcfc4328e5d69431221cc73992df3f4\": container with ID starting with 80a73e88fd8eb4d9f137038a5a6508422dcfc4328e5d69431221cc73992df3f4 not found: ID does not exist" Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.903765 4733 scope.go:117] "RemoveContainer" containerID="017614a4cc5c9d894b40b309a350ff8d25e8495d0c82dcb2a7ab58c0802f74dc" Feb 02 15:37:27 crc kubenswrapper[4733]: E0202 15:37:27.917776 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"017614a4cc5c9d894b40b309a350ff8d25e8495d0c82dcb2a7ab58c0802f74dc\": container with ID starting with 017614a4cc5c9d894b40b309a350ff8d25e8495d0c82dcb2a7ab58c0802f74dc not found: ID does not exist" containerID="017614a4cc5c9d894b40b309a350ff8d25e8495d0c82dcb2a7ab58c0802f74dc" Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.917822 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"017614a4cc5c9d894b40b309a350ff8d25e8495d0c82dcb2a7ab58c0802f74dc"} err="failed to get container status \"017614a4cc5c9d894b40b309a350ff8d25e8495d0c82dcb2a7ab58c0802f74dc\": rpc error: code = NotFound desc = could not find container \"017614a4cc5c9d894b40b309a350ff8d25e8495d0c82dcb2a7ab58c0802f74dc\": container with ID starting with 017614a4cc5c9d894b40b309a350ff8d25e8495d0c82dcb2a7ab58c0802f74dc not found: ID does not exist" Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.917848 4733 scope.go:117] "RemoveContainer" containerID="80a73e88fd8eb4d9f137038a5a6508422dcfc4328e5d69431221cc73992df3f4" Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.918765 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80a73e88fd8eb4d9f137038a5a6508422dcfc4328e5d69431221cc73992df3f4"} err="failed to get container status \"80a73e88fd8eb4d9f137038a5a6508422dcfc4328e5d69431221cc73992df3f4\": rpc error: code = NotFound desc = could not find container \"80a73e88fd8eb4d9f137038a5a6508422dcfc4328e5d69431221cc73992df3f4\": container with ID starting with 80a73e88fd8eb4d9f137038a5a6508422dcfc4328e5d69431221cc73992df3f4 not found: ID does not exist" Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.918808 4733 scope.go:117] "RemoveContainer" containerID="017614a4cc5c9d894b40b309a350ff8d25e8495d0c82dcb2a7ab58c0802f74dc" Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.919295 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"017614a4cc5c9d894b40b309a350ff8d25e8495d0c82dcb2a7ab58c0802f74dc"} err="failed to get container status \"017614a4cc5c9d894b40b309a350ff8d25e8495d0c82dcb2a7ab58c0802f74dc\": rpc error: code = NotFound desc = could not find container \"017614a4cc5c9d894b40b309a350ff8d25e8495d0c82dcb2a7ab58c0802f74dc\": container with ID starting with 017614a4cc5c9d894b40b309a350ff8d25e8495d0c82dcb2a7ab58c0802f74dc not found: ID does not exist" Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.924654 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf8aaefd-448c-4b30-beb7-6590b35a5209-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bf8aaefd-448c-4b30-beb7-6590b35a5209" (UID: "bf8aaefd-448c-4b30-beb7-6590b35a5209"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.961822 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf8aaefd-448c-4b30-beb7-6590b35a5209-config-data" (OuterVolumeSpecName: "config-data") pod "bf8aaefd-448c-4b30-beb7-6590b35a5209" (UID: "bf8aaefd-448c-4b30-beb7-6590b35a5209"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.983679 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf8aaefd-448c-4b30-beb7-6590b35a5209-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.983712 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf8aaefd-448c-4b30-beb7-6590b35a5209-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.983728 4733 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bf8aaefd-448c-4b30-beb7-6590b35a5209-ceph\") on node \"crc\" DevicePath \"\"" Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.983739 4733 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bf8aaefd-448c-4b30-beb7-6590b35a5209-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.983751 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zmqv9\" (UniqueName: \"kubernetes.io/projected/bf8aaefd-448c-4b30-beb7-6590b35a5209-kube-api-access-zmqv9\") on node \"crc\" DevicePath \"\"" Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.983763 4733 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bf8aaefd-448c-4b30-beb7-6590b35a5209-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.983775 4733 reconciler_common.go:293] "Volume detached for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/bf8aaefd-448c-4b30-beb7-6590b35a5209-var-lib-manila\") on node \"crc\" DevicePath \"\"" Feb 02 15:37:27 crc kubenswrapper[4733]: I0202 15:37:27.983787 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf8aaefd-448c-4b30-beb7-6590b35a5209-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 15:37:28 crc kubenswrapper[4733]: I0202 15:37:28.135189 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/manila-share-share0-0"] Feb 02 15:37:28 crc kubenswrapper[4733]: I0202 15:37:28.139791 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["manila-kuttl-tests/manila-share-share0-0"] Feb 02 15:37:28 crc kubenswrapper[4733]: I0202 15:37:28.821818 4733 generic.go:334] "Generic (PLEG): container finished" podID="27a3df30-36df-4380-b693-18775d6b573a" containerID="7909cf6a610bf1ebc516cb20c801752e47646be453177b9151b29750fcba7c66" exitCode=0 Feb 02 15:37:28 crc kubenswrapper[4733]: I0202 15:37:28.821991 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manilaeb2d-account-delete-kx6cd" event={"ID":"27a3df30-36df-4380-b693-18775d6b573a","Type":"ContainerDied","Data":"7909cf6a610bf1ebc516cb20c801752e47646be453177b9151b29750fcba7c66"} Feb 02 15:37:29 crc kubenswrapper[4733]: I0202 15:37:29.263622 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf8aaefd-448c-4b30-beb7-6590b35a5209" path="/var/lib/kubelet/pods/bf8aaefd-448c-4b30-beb7-6590b35a5209/volumes" Feb 02 15:37:30 crc kubenswrapper[4733]: I0202 15:37:30.131531 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manilaeb2d-account-delete-kx6cd" Feb 02 15:37:30 crc kubenswrapper[4733]: I0202 15:37:30.213612 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ddcws\" (UniqueName: \"kubernetes.io/projected/27a3df30-36df-4380-b693-18775d6b573a-kube-api-access-ddcws\") pod \"27a3df30-36df-4380-b693-18775d6b573a\" (UID: \"27a3df30-36df-4380-b693-18775d6b573a\") " Feb 02 15:37:30 crc kubenswrapper[4733]: I0202 15:37:30.213677 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/27a3df30-36df-4380-b693-18775d6b573a-operator-scripts\") pod \"27a3df30-36df-4380-b693-18775d6b573a\" (UID: \"27a3df30-36df-4380-b693-18775d6b573a\") " Feb 02 15:37:30 crc kubenswrapper[4733]: I0202 15:37:30.215975 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27a3df30-36df-4380-b693-18775d6b573a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "27a3df30-36df-4380-b693-18775d6b573a" (UID: "27a3df30-36df-4380-b693-18775d6b573a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:37:30 crc kubenswrapper[4733]: I0202 15:37:30.240107 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27a3df30-36df-4380-b693-18775d6b573a-kube-api-access-ddcws" (OuterVolumeSpecName: "kube-api-access-ddcws") pod "27a3df30-36df-4380-b693-18775d6b573a" (UID: "27a3df30-36df-4380-b693-18775d6b573a"). InnerVolumeSpecName "kube-api-access-ddcws". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:37:30 crc kubenswrapper[4733]: I0202 15:37:30.315378 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ddcws\" (UniqueName: \"kubernetes.io/projected/27a3df30-36df-4380-b693-18775d6b573a-kube-api-access-ddcws\") on node \"crc\" DevicePath \"\"" Feb 02 15:37:30 crc kubenswrapper[4733]: I0202 15:37:30.315416 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/27a3df30-36df-4380-b693-18775d6b573a-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 15:37:30 crc kubenswrapper[4733]: I0202 15:37:30.515227 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-api-0" Feb 02 15:37:30 crc kubenswrapper[4733]: I0202 15:37:30.618706 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4-scripts\") pod \"f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4\" (UID: \"f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4\") " Feb 02 15:37:30 crc kubenswrapper[4733]: I0202 15:37:30.619052 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4-config-data-custom\") pod \"f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4\" (UID: \"f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4\") " Feb 02 15:37:30 crc kubenswrapper[4733]: I0202 15:37:30.619119 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hzm8n\" (UniqueName: \"kubernetes.io/projected/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4-kube-api-access-hzm8n\") pod \"f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4\" (UID: \"f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4\") " Feb 02 15:37:30 crc kubenswrapper[4733]: I0202 15:37:30.619151 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4-etc-machine-id\") pod \"f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4\" (UID: \"f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4\") " Feb 02 15:37:30 crc kubenswrapper[4733]: I0202 15:37:30.619197 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4-combined-ca-bundle\") pod \"f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4\" (UID: \"f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4\") " Feb 02 15:37:30 crc kubenswrapper[4733]: I0202 15:37:30.619255 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4-internal-tls-certs\") pod \"f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4\" (UID: \"f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4\") " Feb 02 15:37:30 crc kubenswrapper[4733]: I0202 15:37:30.619283 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4-config-data\") pod \"f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4\" (UID: \"f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4\") " Feb 02 15:37:30 crc kubenswrapper[4733]: I0202 15:37:30.619287 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4" (UID: "f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 15:37:30 crc kubenswrapper[4733]: I0202 15:37:30.619354 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4-logs\") pod \"f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4\" (UID: \"f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4\") " Feb 02 15:37:30 crc kubenswrapper[4733]: I0202 15:37:30.619384 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4-public-tls-certs\") pod \"f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4\" (UID: \"f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4\") " Feb 02 15:37:30 crc kubenswrapper[4733]: I0202 15:37:30.619849 4733 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 02 15:37:30 crc kubenswrapper[4733]: I0202 15:37:30.621419 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4-logs" (OuterVolumeSpecName: "logs") pod "f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4" (UID: "f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:37:30 crc kubenswrapper[4733]: I0202 15:37:30.635733 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4-scripts" (OuterVolumeSpecName: "scripts") pod "f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4" (UID: "f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:37:30 crc kubenswrapper[4733]: I0202 15:37:30.637712 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4" (UID: "f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:37:30 crc kubenswrapper[4733]: I0202 15:37:30.648835 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4" (UID: "f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:37:30 crc kubenswrapper[4733]: I0202 15:37:30.651094 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4-kube-api-access-hzm8n" (OuterVolumeSpecName: "kube-api-access-hzm8n") pod "f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4" (UID: "f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4"). InnerVolumeSpecName "kube-api-access-hzm8n". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:37:30 crc kubenswrapper[4733]: I0202 15:37:30.651750 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4" (UID: "f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:37:30 crc kubenswrapper[4733]: I0202 15:37:30.655661 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4-config-data" (OuterVolumeSpecName: "config-data") pod "f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4" (UID: "f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:37:30 crc kubenswrapper[4733]: I0202 15:37:30.668770 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4" (UID: "f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:37:30 crc kubenswrapper[4733]: I0202 15:37:30.721314 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 15:37:30 crc kubenswrapper[4733]: I0202 15:37:30.721347 4733 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 02 15:37:30 crc kubenswrapper[4733]: I0202 15:37:30.721359 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hzm8n\" (UniqueName: \"kubernetes.io/projected/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4-kube-api-access-hzm8n\") on node \"crc\" DevicePath \"\"" Feb 02 15:37:30 crc kubenswrapper[4733]: I0202 15:37:30.721369 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 15:37:30 crc kubenswrapper[4733]: I0202 15:37:30.721377 4733 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 15:37:30 crc kubenswrapper[4733]: I0202 15:37:30.721387 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 15:37:30 crc kubenswrapper[4733]: I0202 15:37:30.721397 4733 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4-logs\") on node \"crc\" DevicePath \"\"" Feb 02 15:37:30 crc kubenswrapper[4733]: I0202 15:37:30.721404 4733 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 02 15:37:30 crc kubenswrapper[4733]: I0202 15:37:30.852660 4733 generic.go:334] "Generic (PLEG): container finished" podID="f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4" containerID="8e5d40a60ab730b2ebb52872f09f32bf64a8e1c12a566815bdf9eb1eec77d9cb" exitCode=0 Feb 02 15:37:30 crc kubenswrapper[4733]: I0202 15:37:30.852824 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-api-0" Feb 02 15:37:30 crc kubenswrapper[4733]: I0202 15:37:30.855099 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-api-0" event={"ID":"f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4","Type":"ContainerDied","Data":"8e5d40a60ab730b2ebb52872f09f32bf64a8e1c12a566815bdf9eb1eec77d9cb"} Feb 02 15:37:30 crc kubenswrapper[4733]: I0202 15:37:30.855176 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-api-0" event={"ID":"f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4","Type":"ContainerDied","Data":"c745e7444ca635bd5b159a1c60f5ec0c435d991afc72b2586220a0998f7b3c41"} Feb 02 15:37:30 crc kubenswrapper[4733]: I0202 15:37:30.855197 4733 scope.go:117] "RemoveContainer" containerID="8e5d40a60ab730b2ebb52872f09f32bf64a8e1c12a566815bdf9eb1eec77d9cb" Feb 02 15:37:30 crc kubenswrapper[4733]: I0202 15:37:30.859275 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manilaeb2d-account-delete-kx6cd" Feb 02 15:37:30 crc kubenswrapper[4733]: I0202 15:37:30.860780 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manilaeb2d-account-delete-kx6cd" event={"ID":"27a3df30-36df-4380-b693-18775d6b573a","Type":"ContainerDied","Data":"981688b574f254944ebc024b33116050042703ca3a5f6fb05f5a26ad6ed270b5"} Feb 02 15:37:30 crc kubenswrapper[4733]: I0202 15:37:30.860803 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="981688b574f254944ebc024b33116050042703ca3a5f6fb05f5a26ad6ed270b5" Feb 02 15:37:30 crc kubenswrapper[4733]: I0202 15:37:30.862233 4733 generic.go:334] "Generic (PLEG): container finished" podID="eeaf18e2-5cfc-4795-9f47-5c42f217c7cb" containerID="a1c73b28f4ba7af7cc6b7f195dee621cd2d41b88829f92a723252238e31b3ab2" exitCode=0 Feb 02 15:37:30 crc kubenswrapper[4733]: I0202 15:37:30.862257 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-scheduler-0" event={"ID":"eeaf18e2-5cfc-4795-9f47-5c42f217c7cb","Type":"ContainerDied","Data":"a1c73b28f4ba7af7cc6b7f195dee621cd2d41b88829f92a723252238e31b3ab2"} Feb 02 15:37:30 crc kubenswrapper[4733]: I0202 15:37:30.889925 4733 scope.go:117] "RemoveContainer" containerID="b98b8e2b776fbd00e5c51fab9b2d2d57009eaf7e21b3d40d65fa1cbbb522f7d2" Feb 02 15:37:30 crc kubenswrapper[4733]: I0202 15:37:30.908594 4733 scope.go:117] "RemoveContainer" containerID="8e5d40a60ab730b2ebb52872f09f32bf64a8e1c12a566815bdf9eb1eec77d9cb" Feb 02 15:37:30 crc kubenswrapper[4733]: I0202 15:37:30.908684 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/manila-api-0"] Feb 02 15:37:30 crc kubenswrapper[4733]: E0202 15:37:30.909741 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e5d40a60ab730b2ebb52872f09f32bf64a8e1c12a566815bdf9eb1eec77d9cb\": container with ID starting with 8e5d40a60ab730b2ebb52872f09f32bf64a8e1c12a566815bdf9eb1eec77d9cb not found: ID does not exist" containerID="8e5d40a60ab730b2ebb52872f09f32bf64a8e1c12a566815bdf9eb1eec77d9cb" Feb 02 15:37:30 crc kubenswrapper[4733]: I0202 15:37:30.909765 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e5d40a60ab730b2ebb52872f09f32bf64a8e1c12a566815bdf9eb1eec77d9cb"} err="failed to get container status \"8e5d40a60ab730b2ebb52872f09f32bf64a8e1c12a566815bdf9eb1eec77d9cb\": rpc error: code = NotFound desc = could not find container \"8e5d40a60ab730b2ebb52872f09f32bf64a8e1c12a566815bdf9eb1eec77d9cb\": container with ID starting with 8e5d40a60ab730b2ebb52872f09f32bf64a8e1c12a566815bdf9eb1eec77d9cb not found: ID does not exist" Feb 02 15:37:30 crc kubenswrapper[4733]: I0202 15:37:30.909786 4733 scope.go:117] "RemoveContainer" containerID="b98b8e2b776fbd00e5c51fab9b2d2d57009eaf7e21b3d40d65fa1cbbb522f7d2" Feb 02 15:37:30 crc kubenswrapper[4733]: E0202 15:37:30.910602 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b98b8e2b776fbd00e5c51fab9b2d2d57009eaf7e21b3d40d65fa1cbbb522f7d2\": container with ID starting with b98b8e2b776fbd00e5c51fab9b2d2d57009eaf7e21b3d40d65fa1cbbb522f7d2 not found: ID does not exist" containerID="b98b8e2b776fbd00e5c51fab9b2d2d57009eaf7e21b3d40d65fa1cbbb522f7d2" Feb 02 15:37:30 crc kubenswrapper[4733]: I0202 15:37:30.910722 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b98b8e2b776fbd00e5c51fab9b2d2d57009eaf7e21b3d40d65fa1cbbb522f7d2"} err="failed to get container status \"b98b8e2b776fbd00e5c51fab9b2d2d57009eaf7e21b3d40d65fa1cbbb522f7d2\": rpc error: code = NotFound desc = could not find container \"b98b8e2b776fbd00e5c51fab9b2d2d57009eaf7e21b3d40d65fa1cbbb522f7d2\": container with ID starting with b98b8e2b776fbd00e5c51fab9b2d2d57009eaf7e21b3d40d65fa1cbbb522f7d2 not found: ID does not exist" Feb 02 15:37:30 crc kubenswrapper[4733]: I0202 15:37:30.920779 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["manila-kuttl-tests/manila-api-0"] Feb 02 15:37:30 crc kubenswrapper[4733]: I0202 15:37:30.984898 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:37:31 crc kubenswrapper[4733]: I0202 15:37:31.126583 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eeaf18e2-5cfc-4795-9f47-5c42f217c7cb-config-data\") pod \"eeaf18e2-5cfc-4795-9f47-5c42f217c7cb\" (UID: \"eeaf18e2-5cfc-4795-9f47-5c42f217c7cb\") " Feb 02 15:37:31 crc kubenswrapper[4733]: I0202 15:37:31.126666 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eeaf18e2-5cfc-4795-9f47-5c42f217c7cb-config-data-custom\") pod \"eeaf18e2-5cfc-4795-9f47-5c42f217c7cb\" (UID: \"eeaf18e2-5cfc-4795-9f47-5c42f217c7cb\") " Feb 02 15:37:31 crc kubenswrapper[4733]: I0202 15:37:31.126719 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eeaf18e2-5cfc-4795-9f47-5c42f217c7cb-scripts\") pod \"eeaf18e2-5cfc-4795-9f47-5c42f217c7cb\" (UID: \"eeaf18e2-5cfc-4795-9f47-5c42f217c7cb\") " Feb 02 15:37:31 crc kubenswrapper[4733]: I0202 15:37:31.126744 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eeaf18e2-5cfc-4795-9f47-5c42f217c7cb-combined-ca-bundle\") pod \"eeaf18e2-5cfc-4795-9f47-5c42f217c7cb\" (UID: \"eeaf18e2-5cfc-4795-9f47-5c42f217c7cb\") " Feb 02 15:37:31 crc kubenswrapper[4733]: I0202 15:37:31.126790 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dk5gw\" (UniqueName: \"kubernetes.io/projected/eeaf18e2-5cfc-4795-9f47-5c42f217c7cb-kube-api-access-dk5gw\") pod \"eeaf18e2-5cfc-4795-9f47-5c42f217c7cb\" (UID: \"eeaf18e2-5cfc-4795-9f47-5c42f217c7cb\") " Feb 02 15:37:31 crc kubenswrapper[4733]: I0202 15:37:31.126888 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eeaf18e2-5cfc-4795-9f47-5c42f217c7cb-etc-machine-id\") pod \"eeaf18e2-5cfc-4795-9f47-5c42f217c7cb\" (UID: \"eeaf18e2-5cfc-4795-9f47-5c42f217c7cb\") " Feb 02 15:37:31 crc kubenswrapper[4733]: I0202 15:37:31.127118 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eeaf18e2-5cfc-4795-9f47-5c42f217c7cb-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "eeaf18e2-5cfc-4795-9f47-5c42f217c7cb" (UID: "eeaf18e2-5cfc-4795-9f47-5c42f217c7cb"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 15:37:31 crc kubenswrapper[4733]: I0202 15:37:31.127281 4733 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eeaf18e2-5cfc-4795-9f47-5c42f217c7cb-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 02 15:37:31 crc kubenswrapper[4733]: I0202 15:37:31.129859 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eeaf18e2-5cfc-4795-9f47-5c42f217c7cb-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "eeaf18e2-5cfc-4795-9f47-5c42f217c7cb" (UID: "eeaf18e2-5cfc-4795-9f47-5c42f217c7cb"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:37:31 crc kubenswrapper[4733]: I0202 15:37:31.129999 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eeaf18e2-5cfc-4795-9f47-5c42f217c7cb-scripts" (OuterVolumeSpecName: "scripts") pod "eeaf18e2-5cfc-4795-9f47-5c42f217c7cb" (UID: "eeaf18e2-5cfc-4795-9f47-5c42f217c7cb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:37:31 crc kubenswrapper[4733]: I0202 15:37:31.131086 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eeaf18e2-5cfc-4795-9f47-5c42f217c7cb-kube-api-access-dk5gw" (OuterVolumeSpecName: "kube-api-access-dk5gw") pod "eeaf18e2-5cfc-4795-9f47-5c42f217c7cb" (UID: "eeaf18e2-5cfc-4795-9f47-5c42f217c7cb"). InnerVolumeSpecName "kube-api-access-dk5gw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:37:31 crc kubenswrapper[4733]: I0202 15:37:31.175725 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eeaf18e2-5cfc-4795-9f47-5c42f217c7cb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eeaf18e2-5cfc-4795-9f47-5c42f217c7cb" (UID: "eeaf18e2-5cfc-4795-9f47-5c42f217c7cb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:37:31 crc kubenswrapper[4733]: I0202 15:37:31.196935 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eeaf18e2-5cfc-4795-9f47-5c42f217c7cb-config-data" (OuterVolumeSpecName: "config-data") pod "eeaf18e2-5cfc-4795-9f47-5c42f217c7cb" (UID: "eeaf18e2-5cfc-4795-9f47-5c42f217c7cb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:37:31 crc kubenswrapper[4733]: I0202 15:37:31.228444 4733 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eeaf18e2-5cfc-4795-9f47-5c42f217c7cb-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 02 15:37:31 crc kubenswrapper[4733]: I0202 15:37:31.228480 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eeaf18e2-5cfc-4795-9f47-5c42f217c7cb-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 15:37:31 crc kubenswrapper[4733]: I0202 15:37:31.228489 4733 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eeaf18e2-5cfc-4795-9f47-5c42f217c7cb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 02 15:37:31 crc kubenswrapper[4733]: I0202 15:37:31.228498 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dk5gw\" (UniqueName: \"kubernetes.io/projected/eeaf18e2-5cfc-4795-9f47-5c42f217c7cb-kube-api-access-dk5gw\") on node \"crc\" DevicePath \"\"" Feb 02 15:37:31 crc kubenswrapper[4733]: I0202 15:37:31.228511 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eeaf18e2-5cfc-4795-9f47-5c42f217c7cb-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 15:37:31 crc kubenswrapper[4733]: I0202 15:37:31.269080 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4" path="/var/lib/kubelet/pods/f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4/volumes" Feb 02 15:37:31 crc kubenswrapper[4733]: I0202 15:37:31.875442 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:37:31 crc kubenswrapper[4733]: I0202 15:37:31.875829 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-scheduler-0" event={"ID":"eeaf18e2-5cfc-4795-9f47-5c42f217c7cb","Type":"ContainerDied","Data":"187c94bb8a061c783791b8614b7a95fe91a47032fe4be25304c3117dd9430fff"} Feb 02 15:37:31 crc kubenswrapper[4733]: I0202 15:37:31.875873 4733 scope.go:117] "RemoveContainer" containerID="ea5d03fc685f9d6312eb6363d820b95b9da351a7b850efb702d75bd4714f225a" Feb 02 15:37:31 crc kubenswrapper[4733]: I0202 15:37:31.912281 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/manila-db-create-vq79x"] Feb 02 15:37:31 crc kubenswrapper[4733]: I0202 15:37:31.917874 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["manila-kuttl-tests/manila-db-create-vq79x"] Feb 02 15:37:31 crc kubenswrapper[4733]: I0202 15:37:31.926228 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/manila-scheduler-0"] Feb 02 15:37:31 crc kubenswrapper[4733]: I0202 15:37:31.931615 4733 scope.go:117] "RemoveContainer" containerID="a1c73b28f4ba7af7cc6b7f195dee621cd2d41b88829f92a723252238e31b3ab2" Feb 02 15:37:31 crc kubenswrapper[4733]: I0202 15:37:31.931667 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["manila-kuttl-tests/manila-scheduler-0"] Feb 02 15:37:31 crc kubenswrapper[4733]: I0202 15:37:31.938683 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/manila-eb2d-account-create-update-vzhkk"] Feb 02 15:37:31 crc kubenswrapper[4733]: I0202 15:37:31.945218 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/manilaeb2d-account-delete-kx6cd"] Feb 02 15:37:31 crc kubenswrapper[4733]: I0202 15:37:31.951307 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["manila-kuttl-tests/manilaeb2d-account-delete-kx6cd"] Feb 02 15:37:31 crc kubenswrapper[4733]: I0202 15:37:31.955036 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["manila-kuttl-tests/manila-eb2d-account-create-update-vzhkk"] Feb 02 15:37:31 crc kubenswrapper[4733]: I0202 15:37:31.981676 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["manila-kuttl-tests/manila-db-create-jwlhx"] Feb 02 15:37:31 crc kubenswrapper[4733]: E0202 15:37:31.981881 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf8aaefd-448c-4b30-beb7-6590b35a5209" containerName="manila-share" Feb 02 15:37:31 crc kubenswrapper[4733]: I0202 15:37:31.981891 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf8aaefd-448c-4b30-beb7-6590b35a5209" containerName="manila-share" Feb 02 15:37:31 crc kubenswrapper[4733]: E0202 15:37:31.981904 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4" containerName="manila-api-log" Feb 02 15:37:31 crc kubenswrapper[4733]: I0202 15:37:31.981910 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4" containerName="manila-api-log" Feb 02 15:37:31 crc kubenswrapper[4733]: E0202 15:37:31.981921 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eeaf18e2-5cfc-4795-9f47-5c42f217c7cb" containerName="manila-scheduler" Feb 02 15:37:31 crc kubenswrapper[4733]: I0202 15:37:31.981927 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="eeaf18e2-5cfc-4795-9f47-5c42f217c7cb" containerName="manila-scheduler" Feb 02 15:37:31 crc kubenswrapper[4733]: E0202 15:37:31.981941 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eeaf18e2-5cfc-4795-9f47-5c42f217c7cb" containerName="probe" Feb 02 15:37:31 crc kubenswrapper[4733]: I0202 15:37:31.981947 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="eeaf18e2-5cfc-4795-9f47-5c42f217c7cb" containerName="probe" Feb 02 15:37:31 crc kubenswrapper[4733]: E0202 15:37:31.981958 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27a3df30-36df-4380-b693-18775d6b573a" containerName="mariadb-account-delete" Feb 02 15:37:31 crc kubenswrapper[4733]: I0202 15:37:31.981964 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="27a3df30-36df-4380-b693-18775d6b573a" containerName="mariadb-account-delete" Feb 02 15:37:31 crc kubenswrapper[4733]: E0202 15:37:31.981972 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf8aaefd-448c-4b30-beb7-6590b35a5209" containerName="probe" Feb 02 15:37:31 crc kubenswrapper[4733]: I0202 15:37:31.981977 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf8aaefd-448c-4b30-beb7-6590b35a5209" containerName="probe" Feb 02 15:37:31 crc kubenswrapper[4733]: E0202 15:37:31.981988 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4" containerName="manila-api" Feb 02 15:37:31 crc kubenswrapper[4733]: I0202 15:37:31.981993 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4" containerName="manila-api" Feb 02 15:37:31 crc kubenswrapper[4733]: I0202 15:37:31.982088 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="27a3df30-36df-4380-b693-18775d6b573a" containerName="mariadb-account-delete" Feb 02 15:37:31 crc kubenswrapper[4733]: I0202 15:37:31.982099 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="eeaf18e2-5cfc-4795-9f47-5c42f217c7cb" containerName="manila-scheduler" Feb 02 15:37:31 crc kubenswrapper[4733]: I0202 15:37:31.982106 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf8aaefd-448c-4b30-beb7-6590b35a5209" containerName="probe" Feb 02 15:37:31 crc kubenswrapper[4733]: I0202 15:37:31.982117 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4" containerName="manila-api" Feb 02 15:37:31 crc kubenswrapper[4733]: I0202 15:37:31.982127 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf8aaefd-448c-4b30-beb7-6590b35a5209" containerName="manila-share" Feb 02 15:37:31 crc kubenswrapper[4733]: I0202 15:37:31.982135 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="eeaf18e2-5cfc-4795-9f47-5c42f217c7cb" containerName="probe" Feb 02 15:37:31 crc kubenswrapper[4733]: I0202 15:37:31.982143 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8d44cee-b95d-40e0-9cfc-ebf3f7b87db4" containerName="manila-api-log" Feb 02 15:37:31 crc kubenswrapper[4733]: I0202 15:37:31.982535 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-db-create-jwlhx" Feb 02 15:37:31 crc kubenswrapper[4733]: I0202 15:37:31.992252 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/manila-db-create-jwlhx"] Feb 02 15:37:32 crc kubenswrapper[4733]: I0202 15:37:32.090487 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["manila-kuttl-tests/manila-a9fc-account-create-update-9mw4z"] Feb 02 15:37:32 crc kubenswrapper[4733]: I0202 15:37:32.108707 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-a9fc-account-create-update-9mw4z" Feb 02 15:37:32 crc kubenswrapper[4733]: I0202 15:37:32.114258 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"manila-db-secret" Feb 02 15:37:32 crc kubenswrapper[4733]: I0202 15:37:32.133083 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/manila-a9fc-account-create-update-9mw4z"] Feb 02 15:37:32 crc kubenswrapper[4733]: I0202 15:37:32.148124 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6tjz\" (UniqueName: \"kubernetes.io/projected/ef267821-af74-4e8a-bcd6-216f35b43880-kube-api-access-m6tjz\") pod \"manila-db-create-jwlhx\" (UID: \"ef267821-af74-4e8a-bcd6-216f35b43880\") " pod="manila-kuttl-tests/manila-db-create-jwlhx" Feb 02 15:37:32 crc kubenswrapper[4733]: I0202 15:37:32.148273 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef267821-af74-4e8a-bcd6-216f35b43880-operator-scripts\") pod \"manila-db-create-jwlhx\" (UID: \"ef267821-af74-4e8a-bcd6-216f35b43880\") " pod="manila-kuttl-tests/manila-db-create-jwlhx" Feb 02 15:37:32 crc kubenswrapper[4733]: I0202 15:37:32.250038 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6tjz\" (UniqueName: \"kubernetes.io/projected/ef267821-af74-4e8a-bcd6-216f35b43880-kube-api-access-m6tjz\") pod \"manila-db-create-jwlhx\" (UID: \"ef267821-af74-4e8a-bcd6-216f35b43880\") " pod="manila-kuttl-tests/manila-db-create-jwlhx" Feb 02 15:37:32 crc kubenswrapper[4733]: I0202 15:37:32.250099 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whkn4\" (UniqueName: \"kubernetes.io/projected/5b0e265c-62db-493c-9385-863c39c53ed6-kube-api-access-whkn4\") pod \"manila-a9fc-account-create-update-9mw4z\" (UID: \"5b0e265c-62db-493c-9385-863c39c53ed6\") " pod="manila-kuttl-tests/manila-a9fc-account-create-update-9mw4z" Feb 02 15:37:32 crc kubenswrapper[4733]: I0202 15:37:32.250190 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef267821-af74-4e8a-bcd6-216f35b43880-operator-scripts\") pod \"manila-db-create-jwlhx\" (UID: \"ef267821-af74-4e8a-bcd6-216f35b43880\") " pod="manila-kuttl-tests/manila-db-create-jwlhx" Feb 02 15:37:32 crc kubenswrapper[4733]: I0202 15:37:32.250244 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5b0e265c-62db-493c-9385-863c39c53ed6-operator-scripts\") pod \"manila-a9fc-account-create-update-9mw4z\" (UID: \"5b0e265c-62db-493c-9385-863c39c53ed6\") " pod="manila-kuttl-tests/manila-a9fc-account-create-update-9mw4z" Feb 02 15:37:32 crc kubenswrapper[4733]: I0202 15:37:32.255344 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef267821-af74-4e8a-bcd6-216f35b43880-operator-scripts\") pod \"manila-db-create-jwlhx\" (UID: \"ef267821-af74-4e8a-bcd6-216f35b43880\") " pod="manila-kuttl-tests/manila-db-create-jwlhx" Feb 02 15:37:32 crc kubenswrapper[4733]: I0202 15:37:32.286715 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6tjz\" (UniqueName: \"kubernetes.io/projected/ef267821-af74-4e8a-bcd6-216f35b43880-kube-api-access-m6tjz\") pod \"manila-db-create-jwlhx\" (UID: \"ef267821-af74-4e8a-bcd6-216f35b43880\") " pod="manila-kuttl-tests/manila-db-create-jwlhx" Feb 02 15:37:32 crc kubenswrapper[4733]: I0202 15:37:32.298623 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-db-create-jwlhx" Feb 02 15:37:32 crc kubenswrapper[4733]: I0202 15:37:32.352260 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5b0e265c-62db-493c-9385-863c39c53ed6-operator-scripts\") pod \"manila-a9fc-account-create-update-9mw4z\" (UID: \"5b0e265c-62db-493c-9385-863c39c53ed6\") " pod="manila-kuttl-tests/manila-a9fc-account-create-update-9mw4z" Feb 02 15:37:32 crc kubenswrapper[4733]: I0202 15:37:32.352386 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whkn4\" (UniqueName: \"kubernetes.io/projected/5b0e265c-62db-493c-9385-863c39c53ed6-kube-api-access-whkn4\") pod \"manila-a9fc-account-create-update-9mw4z\" (UID: \"5b0e265c-62db-493c-9385-863c39c53ed6\") " pod="manila-kuttl-tests/manila-a9fc-account-create-update-9mw4z" Feb 02 15:37:32 crc kubenswrapper[4733]: I0202 15:37:32.353522 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5b0e265c-62db-493c-9385-863c39c53ed6-operator-scripts\") pod \"manila-a9fc-account-create-update-9mw4z\" (UID: \"5b0e265c-62db-493c-9385-863c39c53ed6\") " pod="manila-kuttl-tests/manila-a9fc-account-create-update-9mw4z" Feb 02 15:37:32 crc kubenswrapper[4733]: I0202 15:37:32.382445 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whkn4\" (UniqueName: \"kubernetes.io/projected/5b0e265c-62db-493c-9385-863c39c53ed6-kube-api-access-whkn4\") pod \"manila-a9fc-account-create-update-9mw4z\" (UID: \"5b0e265c-62db-493c-9385-863c39c53ed6\") " pod="manila-kuttl-tests/manila-a9fc-account-create-update-9mw4z" Feb 02 15:37:32 crc kubenswrapper[4733]: I0202 15:37:32.438355 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-a9fc-account-create-update-9mw4z" Feb 02 15:37:32 crc kubenswrapper[4733]: I0202 15:37:32.796293 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/manila-db-create-jwlhx"] Feb 02 15:37:32 crc kubenswrapper[4733]: I0202 15:37:32.893318 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-db-create-jwlhx" event={"ID":"ef267821-af74-4e8a-bcd6-216f35b43880","Type":"ContainerStarted","Data":"e9edd1d4401ca2e14629f55a7caddea5859ed4997c415cc8721d001a4c09a209"} Feb 02 15:37:32 crc kubenswrapper[4733]: W0202 15:37:32.923466 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b0e265c_62db_493c_9385_863c39c53ed6.slice/crio-3c6425f11a8597771e8b558493b790ba9a2fbb4ceff22b272433d8fdde46d708 WatchSource:0}: Error finding container 3c6425f11a8597771e8b558493b790ba9a2fbb4ceff22b272433d8fdde46d708: Status 404 returned error can't find the container with id 3c6425f11a8597771e8b558493b790ba9a2fbb4ceff22b272433d8fdde46d708 Feb 02 15:37:32 crc kubenswrapper[4733]: I0202 15:37:32.924105 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/manila-a9fc-account-create-update-9mw4z"] Feb 02 15:37:33 crc kubenswrapper[4733]: I0202 15:37:33.266523 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2761761e-fb0e-43f8-8ad9-5835b5f7e619" path="/var/lib/kubelet/pods/2761761e-fb0e-43f8-8ad9-5835b5f7e619/volumes" Feb 02 15:37:33 crc kubenswrapper[4733]: I0202 15:37:33.268450 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27a3df30-36df-4380-b693-18775d6b573a" path="/var/lib/kubelet/pods/27a3df30-36df-4380-b693-18775d6b573a/volumes" Feb 02 15:37:33 crc kubenswrapper[4733]: I0202 15:37:33.269635 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="594f5269-2f23-4b68-8339-3871e122a649" path="/var/lib/kubelet/pods/594f5269-2f23-4b68-8339-3871e122a649/volumes" Feb 02 15:37:33 crc kubenswrapper[4733]: I0202 15:37:33.270636 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eeaf18e2-5cfc-4795-9f47-5c42f217c7cb" path="/var/lib/kubelet/pods/eeaf18e2-5cfc-4795-9f47-5c42f217c7cb/volumes" Feb 02 15:37:33 crc kubenswrapper[4733]: I0202 15:37:33.908871 4733 generic.go:334] "Generic (PLEG): container finished" podID="5b0e265c-62db-493c-9385-863c39c53ed6" containerID="461dca89c3c6f78c21b9d2e862cb3b9cef70af44ccef57b4d7e23e7dfa2cde43" exitCode=0 Feb 02 15:37:33 crc kubenswrapper[4733]: I0202 15:37:33.908947 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-a9fc-account-create-update-9mw4z" event={"ID":"5b0e265c-62db-493c-9385-863c39c53ed6","Type":"ContainerDied","Data":"461dca89c3c6f78c21b9d2e862cb3b9cef70af44ccef57b4d7e23e7dfa2cde43"} Feb 02 15:37:33 crc kubenswrapper[4733]: I0202 15:37:33.908975 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-a9fc-account-create-update-9mw4z" event={"ID":"5b0e265c-62db-493c-9385-863c39c53ed6","Type":"ContainerStarted","Data":"3c6425f11a8597771e8b558493b790ba9a2fbb4ceff22b272433d8fdde46d708"} Feb 02 15:37:33 crc kubenswrapper[4733]: I0202 15:37:33.911953 4733 generic.go:334] "Generic (PLEG): container finished" podID="ef267821-af74-4e8a-bcd6-216f35b43880" containerID="57cd1e3b7613fed348ef3bc76d834b92b9b0a397c1157db8837ddab596c5bcad" exitCode=0 Feb 02 15:37:33 crc kubenswrapper[4733]: I0202 15:37:33.911982 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-db-create-jwlhx" event={"ID":"ef267821-af74-4e8a-bcd6-216f35b43880","Type":"ContainerDied","Data":"57cd1e3b7613fed348ef3bc76d834b92b9b0a397c1157db8837ddab596c5bcad"} Feb 02 15:37:35 crc kubenswrapper[4733]: I0202 15:37:35.263949 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-a9fc-account-create-update-9mw4z" Feb 02 15:37:35 crc kubenswrapper[4733]: I0202 15:37:35.268896 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-db-create-jwlhx" Feb 02 15:37:35 crc kubenswrapper[4733]: I0202 15:37:35.400316 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef267821-af74-4e8a-bcd6-216f35b43880-operator-scripts\") pod \"ef267821-af74-4e8a-bcd6-216f35b43880\" (UID: \"ef267821-af74-4e8a-bcd6-216f35b43880\") " Feb 02 15:37:35 crc kubenswrapper[4733]: I0202 15:37:35.400394 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5b0e265c-62db-493c-9385-863c39c53ed6-operator-scripts\") pod \"5b0e265c-62db-493c-9385-863c39c53ed6\" (UID: \"5b0e265c-62db-493c-9385-863c39c53ed6\") " Feb 02 15:37:35 crc kubenswrapper[4733]: I0202 15:37:35.400466 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-whkn4\" (UniqueName: \"kubernetes.io/projected/5b0e265c-62db-493c-9385-863c39c53ed6-kube-api-access-whkn4\") pod \"5b0e265c-62db-493c-9385-863c39c53ed6\" (UID: \"5b0e265c-62db-493c-9385-863c39c53ed6\") " Feb 02 15:37:35 crc kubenswrapper[4733]: I0202 15:37:35.400535 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m6tjz\" (UniqueName: \"kubernetes.io/projected/ef267821-af74-4e8a-bcd6-216f35b43880-kube-api-access-m6tjz\") pod \"ef267821-af74-4e8a-bcd6-216f35b43880\" (UID: \"ef267821-af74-4e8a-bcd6-216f35b43880\") " Feb 02 15:37:35 crc kubenswrapper[4733]: I0202 15:37:35.401402 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef267821-af74-4e8a-bcd6-216f35b43880-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ef267821-af74-4e8a-bcd6-216f35b43880" (UID: "ef267821-af74-4e8a-bcd6-216f35b43880"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:37:35 crc kubenswrapper[4733]: I0202 15:37:35.401519 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b0e265c-62db-493c-9385-863c39c53ed6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5b0e265c-62db-493c-9385-863c39c53ed6" (UID: "5b0e265c-62db-493c-9385-863c39c53ed6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:37:35 crc kubenswrapper[4733]: I0202 15:37:35.407555 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef267821-af74-4e8a-bcd6-216f35b43880-kube-api-access-m6tjz" (OuterVolumeSpecName: "kube-api-access-m6tjz") pod "ef267821-af74-4e8a-bcd6-216f35b43880" (UID: "ef267821-af74-4e8a-bcd6-216f35b43880"). InnerVolumeSpecName "kube-api-access-m6tjz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:37:35 crc kubenswrapper[4733]: I0202 15:37:35.418494 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b0e265c-62db-493c-9385-863c39c53ed6-kube-api-access-whkn4" (OuterVolumeSpecName: "kube-api-access-whkn4") pod "5b0e265c-62db-493c-9385-863c39c53ed6" (UID: "5b0e265c-62db-493c-9385-863c39c53ed6"). InnerVolumeSpecName "kube-api-access-whkn4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:37:35 crc kubenswrapper[4733]: I0202 15:37:35.502027 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef267821-af74-4e8a-bcd6-216f35b43880-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 15:37:35 crc kubenswrapper[4733]: I0202 15:37:35.502075 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5b0e265c-62db-493c-9385-863c39c53ed6-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 15:37:35 crc kubenswrapper[4733]: I0202 15:37:35.502085 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-whkn4\" (UniqueName: \"kubernetes.io/projected/5b0e265c-62db-493c-9385-863c39c53ed6-kube-api-access-whkn4\") on node \"crc\" DevicePath \"\"" Feb 02 15:37:35 crc kubenswrapper[4733]: I0202 15:37:35.502098 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m6tjz\" (UniqueName: \"kubernetes.io/projected/ef267821-af74-4e8a-bcd6-216f35b43880-kube-api-access-m6tjz\") on node \"crc\" DevicePath \"\"" Feb 02 15:37:35 crc kubenswrapper[4733]: I0202 15:37:35.933103 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-a9fc-account-create-update-9mw4z" Feb 02 15:37:35 crc kubenswrapper[4733]: I0202 15:37:35.933159 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-a9fc-account-create-update-9mw4z" event={"ID":"5b0e265c-62db-493c-9385-863c39c53ed6","Type":"ContainerDied","Data":"3c6425f11a8597771e8b558493b790ba9a2fbb4ceff22b272433d8fdde46d708"} Feb 02 15:37:35 crc kubenswrapper[4733]: I0202 15:37:35.933253 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3c6425f11a8597771e8b558493b790ba9a2fbb4ceff22b272433d8fdde46d708" Feb 02 15:37:35 crc kubenswrapper[4733]: I0202 15:37:35.935221 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-db-create-jwlhx" event={"ID":"ef267821-af74-4e8a-bcd6-216f35b43880","Type":"ContainerDied","Data":"e9edd1d4401ca2e14629f55a7caddea5859ed4997c415cc8721d001a4c09a209"} Feb 02 15:37:35 crc kubenswrapper[4733]: I0202 15:37:35.935273 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e9edd1d4401ca2e14629f55a7caddea5859ed4997c415cc8721d001a4c09a209" Feb 02 15:37:35 crc kubenswrapper[4733]: I0202 15:37:35.935339 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-db-create-jwlhx" Feb 02 15:37:37 crc kubenswrapper[4733]: I0202 15:37:37.321743 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["manila-kuttl-tests/manila-db-sync-8q2p6"] Feb 02 15:37:37 crc kubenswrapper[4733]: E0202 15:37:37.322537 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b0e265c-62db-493c-9385-863c39c53ed6" containerName="mariadb-account-create-update" Feb 02 15:37:37 crc kubenswrapper[4733]: I0202 15:37:37.322560 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b0e265c-62db-493c-9385-863c39c53ed6" containerName="mariadb-account-create-update" Feb 02 15:37:37 crc kubenswrapper[4733]: E0202 15:37:37.322604 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef267821-af74-4e8a-bcd6-216f35b43880" containerName="mariadb-database-create" Feb 02 15:37:37 crc kubenswrapper[4733]: I0202 15:37:37.322616 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef267821-af74-4e8a-bcd6-216f35b43880" containerName="mariadb-database-create" Feb 02 15:37:37 crc kubenswrapper[4733]: I0202 15:37:37.322823 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef267821-af74-4e8a-bcd6-216f35b43880" containerName="mariadb-database-create" Feb 02 15:37:37 crc kubenswrapper[4733]: I0202 15:37:37.322853 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b0e265c-62db-493c-9385-863c39c53ed6" containerName="mariadb-account-create-update" Feb 02 15:37:37 crc kubenswrapper[4733]: I0202 15:37:37.323619 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-db-sync-8q2p6" Feb 02 15:37:37 crc kubenswrapper[4733]: I0202 15:37:37.326769 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"manila-config-data" Feb 02 15:37:37 crc kubenswrapper[4733]: I0202 15:37:37.326858 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"manila-manila-dockercfg-mfjs4" Feb 02 15:37:37 crc kubenswrapper[4733]: I0202 15:37:37.331794 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/manila-db-sync-8q2p6"] Feb 02 15:37:37 crc kubenswrapper[4733]: I0202 15:37:37.431889 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzxhx\" (UniqueName: \"kubernetes.io/projected/0e42c981-8832-49a6-b150-811064d2509b-kube-api-access-gzxhx\") pod \"manila-db-sync-8q2p6\" (UID: \"0e42c981-8832-49a6-b150-811064d2509b\") " pod="manila-kuttl-tests/manila-db-sync-8q2p6" Feb 02 15:37:37 crc kubenswrapper[4733]: I0202 15:37:37.432059 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e42c981-8832-49a6-b150-811064d2509b-config-data\") pod \"manila-db-sync-8q2p6\" (UID: \"0e42c981-8832-49a6-b150-811064d2509b\") " pod="manila-kuttl-tests/manila-db-sync-8q2p6" Feb 02 15:37:37 crc kubenswrapper[4733]: I0202 15:37:37.432330 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/0e42c981-8832-49a6-b150-811064d2509b-job-config-data\") pod \"manila-db-sync-8q2p6\" (UID: \"0e42c981-8832-49a6-b150-811064d2509b\") " pod="manila-kuttl-tests/manila-db-sync-8q2p6" Feb 02 15:37:37 crc kubenswrapper[4733]: I0202 15:37:37.533900 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/0e42c981-8832-49a6-b150-811064d2509b-job-config-data\") pod \"manila-db-sync-8q2p6\" (UID: \"0e42c981-8832-49a6-b150-811064d2509b\") " pod="manila-kuttl-tests/manila-db-sync-8q2p6" Feb 02 15:37:37 crc kubenswrapper[4733]: I0202 15:37:37.534003 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzxhx\" (UniqueName: \"kubernetes.io/projected/0e42c981-8832-49a6-b150-811064d2509b-kube-api-access-gzxhx\") pod \"manila-db-sync-8q2p6\" (UID: \"0e42c981-8832-49a6-b150-811064d2509b\") " pod="manila-kuttl-tests/manila-db-sync-8q2p6" Feb 02 15:37:37 crc kubenswrapper[4733]: I0202 15:37:37.534071 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e42c981-8832-49a6-b150-811064d2509b-config-data\") pod \"manila-db-sync-8q2p6\" (UID: \"0e42c981-8832-49a6-b150-811064d2509b\") " pod="manila-kuttl-tests/manila-db-sync-8q2p6" Feb 02 15:37:37 crc kubenswrapper[4733]: I0202 15:37:37.546394 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/0e42c981-8832-49a6-b150-811064d2509b-job-config-data\") pod \"manila-db-sync-8q2p6\" (UID: \"0e42c981-8832-49a6-b150-811064d2509b\") " pod="manila-kuttl-tests/manila-db-sync-8q2p6" Feb 02 15:37:37 crc kubenswrapper[4733]: I0202 15:37:37.547110 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e42c981-8832-49a6-b150-811064d2509b-config-data\") pod \"manila-db-sync-8q2p6\" (UID: \"0e42c981-8832-49a6-b150-811064d2509b\") " pod="manila-kuttl-tests/manila-db-sync-8q2p6" Feb 02 15:37:37 crc kubenswrapper[4733]: I0202 15:37:37.562521 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzxhx\" (UniqueName: \"kubernetes.io/projected/0e42c981-8832-49a6-b150-811064d2509b-kube-api-access-gzxhx\") pod \"manila-db-sync-8q2p6\" (UID: \"0e42c981-8832-49a6-b150-811064d2509b\") " pod="manila-kuttl-tests/manila-db-sync-8q2p6" Feb 02 15:37:37 crc kubenswrapper[4733]: I0202 15:37:37.645712 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-db-sync-8q2p6" Feb 02 15:37:38 crc kubenswrapper[4733]: I0202 15:37:38.144726 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/manila-db-sync-8q2p6"] Feb 02 15:37:38 crc kubenswrapper[4733]: I0202 15:37:38.967318 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-db-sync-8q2p6" event={"ID":"0e42c981-8832-49a6-b150-811064d2509b","Type":"ContainerStarted","Data":"1330407f99c2108f44a26a135750204597d0a5b48f21a356bb3feb3388cf237b"} Feb 02 15:37:38 crc kubenswrapper[4733]: I0202 15:37:38.967373 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-db-sync-8q2p6" event={"ID":"0e42c981-8832-49a6-b150-811064d2509b","Type":"ContainerStarted","Data":"2938f6affb6a036c148bbbd196200d3c51ee79c877ae0d6530e7982e19541382"} Feb 02 15:37:38 crc kubenswrapper[4733]: I0202 15:37:38.988052 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="manila-kuttl-tests/manila-db-sync-8q2p6" podStartSLOduration=1.988030265 podStartE2EDuration="1.988030265s" podCreationTimestamp="2026-02-02 15:37:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:37:38.982513342 +0000 UTC m=+1342.433974710" watchObservedRunningTime="2026-02-02 15:37:38.988030265 +0000 UTC m=+1342.439491643" Feb 02 15:37:39 crc kubenswrapper[4733]: I0202 15:37:39.979159 4733 generic.go:334] "Generic (PLEG): container finished" podID="0e42c981-8832-49a6-b150-811064d2509b" containerID="1330407f99c2108f44a26a135750204597d0a5b48f21a356bb3feb3388cf237b" exitCode=0 Feb 02 15:37:39 crc kubenswrapper[4733]: I0202 15:37:39.979692 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-db-sync-8q2p6" event={"ID":"0e42c981-8832-49a6-b150-811064d2509b","Type":"ContainerDied","Data":"1330407f99c2108f44a26a135750204597d0a5b48f21a356bb3feb3388cf237b"} Feb 02 15:37:41 crc kubenswrapper[4733]: I0202 15:37:41.325952 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-db-sync-8q2p6" Feb 02 15:37:41 crc kubenswrapper[4733]: I0202 15:37:41.403762 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e42c981-8832-49a6-b150-811064d2509b-config-data\") pod \"0e42c981-8832-49a6-b150-811064d2509b\" (UID: \"0e42c981-8832-49a6-b150-811064d2509b\") " Feb 02 15:37:41 crc kubenswrapper[4733]: I0202 15:37:41.404057 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gzxhx\" (UniqueName: \"kubernetes.io/projected/0e42c981-8832-49a6-b150-811064d2509b-kube-api-access-gzxhx\") pod \"0e42c981-8832-49a6-b150-811064d2509b\" (UID: \"0e42c981-8832-49a6-b150-811064d2509b\") " Feb 02 15:37:41 crc kubenswrapper[4733]: I0202 15:37:41.404131 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/0e42c981-8832-49a6-b150-811064d2509b-job-config-data\") pod \"0e42c981-8832-49a6-b150-811064d2509b\" (UID: \"0e42c981-8832-49a6-b150-811064d2509b\") " Feb 02 15:37:41 crc kubenswrapper[4733]: I0202 15:37:41.413298 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e42c981-8832-49a6-b150-811064d2509b-kube-api-access-gzxhx" (OuterVolumeSpecName: "kube-api-access-gzxhx") pod "0e42c981-8832-49a6-b150-811064d2509b" (UID: "0e42c981-8832-49a6-b150-811064d2509b"). InnerVolumeSpecName "kube-api-access-gzxhx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:37:41 crc kubenswrapper[4733]: I0202 15:37:41.415507 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e42c981-8832-49a6-b150-811064d2509b-job-config-data" (OuterVolumeSpecName: "job-config-data") pod "0e42c981-8832-49a6-b150-811064d2509b" (UID: "0e42c981-8832-49a6-b150-811064d2509b"). InnerVolumeSpecName "job-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:37:41 crc kubenswrapper[4733]: I0202 15:37:41.418491 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e42c981-8832-49a6-b150-811064d2509b-config-data" (OuterVolumeSpecName: "config-data") pod "0e42c981-8832-49a6-b150-811064d2509b" (UID: "0e42c981-8832-49a6-b150-811064d2509b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:37:41 crc kubenswrapper[4733]: I0202 15:37:41.506651 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gzxhx\" (UniqueName: \"kubernetes.io/projected/0e42c981-8832-49a6-b150-811064d2509b-kube-api-access-gzxhx\") on node \"crc\" DevicePath \"\"" Feb 02 15:37:41 crc kubenswrapper[4733]: I0202 15:37:41.506918 4733 reconciler_common.go:293] "Volume detached for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/0e42c981-8832-49a6-b150-811064d2509b-job-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 15:37:41 crc kubenswrapper[4733]: I0202 15:37:41.506971 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e42c981-8832-49a6-b150-811064d2509b-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 15:37:41 crc kubenswrapper[4733]: I0202 15:37:41.999237 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-db-sync-8q2p6" event={"ID":"0e42c981-8832-49a6-b150-811064d2509b","Type":"ContainerDied","Data":"2938f6affb6a036c148bbbd196200d3c51ee79c877ae0d6530e7982e19541382"} Feb 02 15:37:41 crc kubenswrapper[4733]: I0202 15:37:41.999584 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2938f6affb6a036c148bbbd196200d3c51ee79c877ae0d6530e7982e19541382" Feb 02 15:37:41 crc kubenswrapper[4733]: I0202 15:37:41.999455 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-db-sync-8q2p6" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.296191 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["manila-kuttl-tests/manila-scheduler-0"] Feb 02 15:37:42 crc kubenswrapper[4733]: E0202 15:37:42.296590 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e42c981-8832-49a6-b150-811064d2509b" containerName="manila-db-sync" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.296621 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e42c981-8832-49a6-b150-811064d2509b" containerName="manila-db-sync" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.296852 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e42c981-8832-49a6-b150-811064d2509b" containerName="manila-db-sync" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.298025 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.301613 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"manila-scripts" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.302132 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"manila-config-data" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.302801 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"manila-scheduler-config-data" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.303112 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"manila-manila-dockercfg-mfjs4" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.371753 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/manila-scheduler-0"] Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.409015 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["manila-kuttl-tests/manila-share-share0-0"] Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.410579 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.415827 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"manila-share-share0-config-data" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.417231 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"ceph-conf-files" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.420095 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9abf907e-85df-4616-a830-56fae1ae3f69-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"9abf907e-85df-4616-a830-56fae1ae3f69\") " pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.420211 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9abf907e-85df-4616-a830-56fae1ae3f69-scripts\") pod \"manila-scheduler-0\" (UID: \"9abf907e-85df-4616-a830-56fae1ae3f69\") " pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.420426 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9abf907e-85df-4616-a830-56fae1ae3f69-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"9abf907e-85df-4616-a830-56fae1ae3f69\") " pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.420466 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9abf907e-85df-4616-a830-56fae1ae3f69-config-data\") pod \"manila-scheduler-0\" (UID: \"9abf907e-85df-4616-a830-56fae1ae3f69\") " pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.420571 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8gmm6\" (UniqueName: \"kubernetes.io/projected/9abf907e-85df-4616-a830-56fae1ae3f69-kube-api-access-8gmm6\") pod \"manila-scheduler-0\" (UID: \"9abf907e-85df-4616-a830-56fae1ae3f69\") " pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.423929 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/manila-share-share0-0"] Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.511893 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["manila-kuttl-tests/manila-api-0"] Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.512869 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-api-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.515046 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"manila-api-config-data" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.521968 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2267a854-0257-4ab1-8951-b82dc2b7350f-config-data\") pod \"manila-share-share0-0\" (UID: \"2267a854-0257-4ab1-8951-b82dc2b7350f\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.522010 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2267a854-0257-4ab1-8951-b82dc2b7350f-etc-machine-id\") pod \"manila-share-share0-0\" (UID: \"2267a854-0257-4ab1-8951-b82dc2b7350f\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.522029 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/2267a854-0257-4ab1-8951-b82dc2b7350f-var-lib-manila\") pod \"manila-share-share0-0\" (UID: \"2267a854-0257-4ab1-8951-b82dc2b7350f\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.522047 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2267a854-0257-4ab1-8951-b82dc2b7350f-ceph\") pod \"manila-share-share0-0\" (UID: \"2267a854-0257-4ab1-8951-b82dc2b7350f\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.522064 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52l96\" (UniqueName: \"kubernetes.io/projected/2267a854-0257-4ab1-8951-b82dc2b7350f-kube-api-access-52l96\") pod \"manila-share-share0-0\" (UID: \"2267a854-0257-4ab1-8951-b82dc2b7350f\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.522240 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9abf907e-85df-4616-a830-56fae1ae3f69-scripts\") pod \"manila-scheduler-0\" (UID: \"9abf907e-85df-4616-a830-56fae1ae3f69\") " pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.522283 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2267a854-0257-4ab1-8951-b82dc2b7350f-config-data-custom\") pod \"manila-share-share0-0\" (UID: \"2267a854-0257-4ab1-8951-b82dc2b7350f\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.522370 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9abf907e-85df-4616-a830-56fae1ae3f69-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"9abf907e-85df-4616-a830-56fae1ae3f69\") " pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.522404 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9abf907e-85df-4616-a830-56fae1ae3f69-config-data\") pod \"manila-scheduler-0\" (UID: \"9abf907e-85df-4616-a830-56fae1ae3f69\") " pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.522448 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2267a854-0257-4ab1-8951-b82dc2b7350f-scripts\") pod \"manila-share-share0-0\" (UID: \"2267a854-0257-4ab1-8951-b82dc2b7350f\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.522524 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8gmm6\" (UniqueName: \"kubernetes.io/projected/9abf907e-85df-4616-a830-56fae1ae3f69-kube-api-access-8gmm6\") pod \"manila-scheduler-0\" (UID: \"9abf907e-85df-4616-a830-56fae1ae3f69\") " pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.522602 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9abf907e-85df-4616-a830-56fae1ae3f69-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"9abf907e-85df-4616-a830-56fae1ae3f69\") " pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.522716 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9abf907e-85df-4616-a830-56fae1ae3f69-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"9abf907e-85df-4616-a830-56fae1ae3f69\") " pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.526429 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/manila-api-0"] Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.526919 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9abf907e-85df-4616-a830-56fae1ae3f69-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"9abf907e-85df-4616-a830-56fae1ae3f69\") " pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.527532 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9abf907e-85df-4616-a830-56fae1ae3f69-scripts\") pod \"manila-scheduler-0\" (UID: \"9abf907e-85df-4616-a830-56fae1ae3f69\") " pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.536878 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9abf907e-85df-4616-a830-56fae1ae3f69-config-data\") pod \"manila-scheduler-0\" (UID: \"9abf907e-85df-4616-a830-56fae1ae3f69\") " pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.557430 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8gmm6\" (UniqueName: \"kubernetes.io/projected/9abf907e-85df-4616-a830-56fae1ae3f69-kube-api-access-8gmm6\") pod \"manila-scheduler-0\" (UID: \"9abf907e-85df-4616-a830-56fae1ae3f69\") " pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.623579 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2267a854-0257-4ab1-8951-b82dc2b7350f-scripts\") pod \"manila-share-share0-0\" (UID: \"2267a854-0257-4ab1-8951-b82dc2b7350f\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.623938 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmmmb\" (UniqueName: \"kubernetes.io/projected/fe12e7c9-118b-4413-8fa4-96745802310e-kube-api-access-dmmmb\") pod \"manila-api-0\" (UID: \"fe12e7c9-118b-4413-8fa4-96745802310e\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.623967 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe12e7c9-118b-4413-8fa4-96745802310e-scripts\") pod \"manila-api-0\" (UID: \"fe12e7c9-118b-4413-8fa4-96745802310e\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.623985 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fe12e7c9-118b-4413-8fa4-96745802310e-logs\") pod \"manila-api-0\" (UID: \"fe12e7c9-118b-4413-8fa4-96745802310e\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.624017 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fe12e7c9-118b-4413-8fa4-96745802310e-etc-machine-id\") pod \"manila-api-0\" (UID: \"fe12e7c9-118b-4413-8fa4-96745802310e\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.624056 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2267a854-0257-4ab1-8951-b82dc2b7350f-config-data\") pod \"manila-share-share0-0\" (UID: \"2267a854-0257-4ab1-8951-b82dc2b7350f\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.624134 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2267a854-0257-4ab1-8951-b82dc2b7350f-etc-machine-id\") pod \"manila-share-share0-0\" (UID: \"2267a854-0257-4ab1-8951-b82dc2b7350f\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.624208 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2267a854-0257-4ab1-8951-b82dc2b7350f-etc-machine-id\") pod \"manila-share-share0-0\" (UID: \"2267a854-0257-4ab1-8951-b82dc2b7350f\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.624150 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/2267a854-0257-4ab1-8951-b82dc2b7350f-var-lib-manila\") pod \"manila-share-share0-0\" (UID: \"2267a854-0257-4ab1-8951-b82dc2b7350f\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.624255 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2267a854-0257-4ab1-8951-b82dc2b7350f-ceph\") pod \"manila-share-share0-0\" (UID: \"2267a854-0257-4ab1-8951-b82dc2b7350f\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.624339 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/2267a854-0257-4ab1-8951-b82dc2b7350f-var-lib-manila\") pod \"manila-share-share0-0\" (UID: \"2267a854-0257-4ab1-8951-b82dc2b7350f\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.624274 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52l96\" (UniqueName: \"kubernetes.io/projected/2267a854-0257-4ab1-8951-b82dc2b7350f-kube-api-access-52l96\") pod \"manila-share-share0-0\" (UID: \"2267a854-0257-4ab1-8951-b82dc2b7350f\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.624459 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2267a854-0257-4ab1-8951-b82dc2b7350f-config-data-custom\") pod \"manila-share-share0-0\" (UID: \"2267a854-0257-4ab1-8951-b82dc2b7350f\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.624532 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fe12e7c9-118b-4413-8fa4-96745802310e-config-data-custom\") pod \"manila-api-0\" (UID: \"fe12e7c9-118b-4413-8fa4-96745802310e\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.624588 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe12e7c9-118b-4413-8fa4-96745802310e-config-data\") pod \"manila-api-0\" (UID: \"fe12e7c9-118b-4413-8fa4-96745802310e\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.626848 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2267a854-0257-4ab1-8951-b82dc2b7350f-scripts\") pod \"manila-share-share0-0\" (UID: \"2267a854-0257-4ab1-8951-b82dc2b7350f\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.628589 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2267a854-0257-4ab1-8951-b82dc2b7350f-config-data\") pod \"manila-share-share0-0\" (UID: \"2267a854-0257-4ab1-8951-b82dc2b7350f\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.628942 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2267a854-0257-4ab1-8951-b82dc2b7350f-config-data-custom\") pod \"manila-share-share0-0\" (UID: \"2267a854-0257-4ab1-8951-b82dc2b7350f\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.629121 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2267a854-0257-4ab1-8951-b82dc2b7350f-ceph\") pod \"manila-share-share0-0\" (UID: \"2267a854-0257-4ab1-8951-b82dc2b7350f\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.633213 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.650885 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52l96\" (UniqueName: \"kubernetes.io/projected/2267a854-0257-4ab1-8951-b82dc2b7350f-kube-api-access-52l96\") pod \"manila-share-share0-0\" (UID: \"2267a854-0257-4ab1-8951-b82dc2b7350f\") " pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.726418 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe12e7c9-118b-4413-8fa4-96745802310e-config-data\") pod \"manila-api-0\" (UID: \"fe12e7c9-118b-4413-8fa4-96745802310e\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.726505 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmmmb\" (UniqueName: \"kubernetes.io/projected/fe12e7c9-118b-4413-8fa4-96745802310e-kube-api-access-dmmmb\") pod \"manila-api-0\" (UID: \"fe12e7c9-118b-4413-8fa4-96745802310e\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.726561 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe12e7c9-118b-4413-8fa4-96745802310e-scripts\") pod \"manila-api-0\" (UID: \"fe12e7c9-118b-4413-8fa4-96745802310e\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.726582 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fe12e7c9-118b-4413-8fa4-96745802310e-logs\") pod \"manila-api-0\" (UID: \"fe12e7c9-118b-4413-8fa4-96745802310e\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.726649 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fe12e7c9-118b-4413-8fa4-96745802310e-etc-machine-id\") pod \"manila-api-0\" (UID: \"fe12e7c9-118b-4413-8fa4-96745802310e\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.726768 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fe12e7c9-118b-4413-8fa4-96745802310e-config-data-custom\") pod \"manila-api-0\" (UID: \"fe12e7c9-118b-4413-8fa4-96745802310e\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.729732 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fe12e7c9-118b-4413-8fa4-96745802310e-logs\") pod \"manila-api-0\" (UID: \"fe12e7c9-118b-4413-8fa4-96745802310e\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.730327 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fe12e7c9-118b-4413-8fa4-96745802310e-etc-machine-id\") pod \"manila-api-0\" (UID: \"fe12e7c9-118b-4413-8fa4-96745802310e\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.730505 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.734837 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe12e7c9-118b-4413-8fa4-96745802310e-scripts\") pod \"manila-api-0\" (UID: \"fe12e7c9-118b-4413-8fa4-96745802310e\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.735376 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe12e7c9-118b-4413-8fa4-96745802310e-config-data\") pod \"manila-api-0\" (UID: \"fe12e7c9-118b-4413-8fa4-96745802310e\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.735495 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fe12e7c9-118b-4413-8fa4-96745802310e-config-data-custom\") pod \"manila-api-0\" (UID: \"fe12e7c9-118b-4413-8fa4-96745802310e\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.753206 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmmmb\" (UniqueName: \"kubernetes.io/projected/fe12e7c9-118b-4413-8fa4-96745802310e-kube-api-access-dmmmb\") pod \"manila-api-0\" (UID: \"fe12e7c9-118b-4413-8fa4-96745802310e\") " pod="manila-kuttl-tests/manila-api-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.827329 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-api-0" Feb 02 15:37:42 crc kubenswrapper[4733]: I0202 15:37:42.989945 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/manila-share-share0-0"] Feb 02 15:37:42 crc kubenswrapper[4733]: W0202 15:37:42.998069 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2267a854_0257_4ab1_8951_b82dc2b7350f.slice/crio-289c652f6d515b9cbad9ecb79e7fa768638c8017c5229ce454d853c166fc18a0 WatchSource:0}: Error finding container 289c652f6d515b9cbad9ecb79e7fa768638c8017c5229ce454d853c166fc18a0: Status 404 returned error can't find the container with id 289c652f6d515b9cbad9ecb79e7fa768638c8017c5229ce454d853c166fc18a0 Feb 02 15:37:43 crc kubenswrapper[4733]: I0202 15:37:43.019075 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-share-share0-0" event={"ID":"2267a854-0257-4ab1-8951-b82dc2b7350f","Type":"ContainerStarted","Data":"289c652f6d515b9cbad9ecb79e7fa768638c8017c5229ce454d853c166fc18a0"} Feb 02 15:37:43 crc kubenswrapper[4733]: I0202 15:37:43.063478 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/manila-scheduler-0"] Feb 02 15:37:43 crc kubenswrapper[4733]: W0202 15:37:43.067812 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9abf907e_85df_4616_a830_56fae1ae3f69.slice/crio-55cf3d26df3efaf1b447714e8d36cb6bdcb9c9307b451802531a312133d09ad0 WatchSource:0}: Error finding container 55cf3d26df3efaf1b447714e8d36cb6bdcb9c9307b451802531a312133d09ad0: Status 404 returned error can't find the container with id 55cf3d26df3efaf1b447714e8d36cb6bdcb9c9307b451802531a312133d09ad0 Feb 02 15:37:43 crc kubenswrapper[4733]: I0202 15:37:43.270803 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/manila-api-0"] Feb 02 15:37:43 crc kubenswrapper[4733]: W0202 15:37:43.286254 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfe12e7c9_118b_4413_8fa4_96745802310e.slice/crio-4bae261d9bc30895052af22344146c6928e78f18b17985b7a57f92c4f560e7d3 WatchSource:0}: Error finding container 4bae261d9bc30895052af22344146c6928e78f18b17985b7a57f92c4f560e7d3: Status 404 returned error can't find the container with id 4bae261d9bc30895052af22344146c6928e78f18b17985b7a57f92c4f560e7d3 Feb 02 15:37:44 crc kubenswrapper[4733]: I0202 15:37:44.027948 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-share-share0-0" event={"ID":"2267a854-0257-4ab1-8951-b82dc2b7350f","Type":"ContainerStarted","Data":"8be19e6ccae7070cbdb2e63b741acd61dd3e9c876427ca89a3f9e49547ab4db1"} Feb 02 15:37:44 crc kubenswrapper[4733]: I0202 15:37:44.029119 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-share-share0-0" event={"ID":"2267a854-0257-4ab1-8951-b82dc2b7350f","Type":"ContainerStarted","Data":"8127c0a93057ef0939fd26da8e2f58eadf4e6ae1f51f0bfb3942c07ad85a5eed"} Feb 02 15:37:44 crc kubenswrapper[4733]: I0202 15:37:44.032350 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-api-0" event={"ID":"fe12e7c9-118b-4413-8fa4-96745802310e","Type":"ContainerStarted","Data":"11acc8fca457138b1b74f4ee6a9396a9e163b6a49c60cad0a5e39487d7377acf"} Feb 02 15:37:44 crc kubenswrapper[4733]: I0202 15:37:44.032389 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-api-0" event={"ID":"fe12e7c9-118b-4413-8fa4-96745802310e","Type":"ContainerStarted","Data":"4bae261d9bc30895052af22344146c6928e78f18b17985b7a57f92c4f560e7d3"} Feb 02 15:37:44 crc kubenswrapper[4733]: I0202 15:37:44.033617 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-scheduler-0" event={"ID":"9abf907e-85df-4616-a830-56fae1ae3f69","Type":"ContainerStarted","Data":"b7e17883ce2314948330a91af64d17c2235f67db1c6a7c62d5c1097a7e2be57f"} Feb 02 15:37:44 crc kubenswrapper[4733]: I0202 15:37:44.033643 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-scheduler-0" event={"ID":"9abf907e-85df-4616-a830-56fae1ae3f69","Type":"ContainerStarted","Data":"1b4db59339def5c0775f38a97cbcb2c76ad7e0dde03ca35dd3b89b993012719d"} Feb 02 15:37:44 crc kubenswrapper[4733]: I0202 15:37:44.033654 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-scheduler-0" event={"ID":"9abf907e-85df-4616-a830-56fae1ae3f69","Type":"ContainerStarted","Data":"55cf3d26df3efaf1b447714e8d36cb6bdcb9c9307b451802531a312133d09ad0"} Feb 02 15:37:44 crc kubenswrapper[4733]: I0202 15:37:44.055143 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="manila-kuttl-tests/manila-share-share0-0" podStartSLOduration=2.055124125 podStartE2EDuration="2.055124125s" podCreationTimestamp="2026-02-02 15:37:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:37:44.051855461 +0000 UTC m=+1347.503316819" watchObservedRunningTime="2026-02-02 15:37:44.055124125 +0000 UTC m=+1347.506585483" Feb 02 15:37:44 crc kubenswrapper[4733]: I0202 15:37:44.080144 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="manila-kuttl-tests/manila-scheduler-0" podStartSLOduration=2.0801225150000002 podStartE2EDuration="2.080122515s" podCreationTimestamp="2026-02-02 15:37:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:37:44.069059988 +0000 UTC m=+1347.520521346" watchObservedRunningTime="2026-02-02 15:37:44.080122515 +0000 UTC m=+1347.531583873" Feb 02 15:37:45 crc kubenswrapper[4733]: I0202 15:37:45.040980 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-api-0" event={"ID":"fe12e7c9-118b-4413-8fa4-96745802310e","Type":"ContainerStarted","Data":"0c0517983e8f69418d96a083e721d09d7185d0fbb013e0b4fc4e93c6e1b2a795"} Feb 02 15:37:45 crc kubenswrapper[4733]: I0202 15:37:45.060980 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="manila-kuttl-tests/manila-api-0" podStartSLOduration=3.060961903 podStartE2EDuration="3.060961903s" podCreationTimestamp="2026-02-02 15:37:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:37:45.058471178 +0000 UTC m=+1348.509932536" watchObservedRunningTime="2026-02-02 15:37:45.060961903 +0000 UTC m=+1348.512423251" Feb 02 15:37:46 crc kubenswrapper[4733]: I0202 15:37:46.045916 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="manila-kuttl-tests/manila-api-0" Feb 02 15:37:52 crc kubenswrapper[4733]: I0202 15:37:52.634077 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:37:52 crc kubenswrapper[4733]: I0202 15:37:52.730885 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:38:04 crc kubenswrapper[4733]: I0202 15:38:04.089783 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:38:04 crc kubenswrapper[4733]: I0202 15:38:04.112572 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="manila-kuttl-tests/manila-api-0" Feb 02 15:38:04 crc kubenswrapper[4733]: I0202 15:38:04.358926 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:38:07 crc kubenswrapper[4733]: I0202 15:38:07.036028 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["manila-kuttl-tests/manila-share-share1-0"] Feb 02 15:38:07 crc kubenswrapper[4733]: I0202 15:38:07.037335 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-share-share1-0" Feb 02 15:38:07 crc kubenswrapper[4733]: I0202 15:38:07.039406 4733 reflector.go:368] Caches populated for *v1.Secret from object-"manila-kuttl-tests"/"manila-share-share1-config-data" Feb 02 15:38:07 crc kubenswrapper[4733]: I0202 15:38:07.052200 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/manila-share-share1-0"] Feb 02 15:38:07 crc kubenswrapper[4733]: I0202 15:38:07.117058 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f33ab3b4-cc56-4046-a8c5-bd859c2df2aa-scripts\") pod \"manila-share-share1-0\" (UID: \"f33ab3b4-cc56-4046-a8c5-bd859c2df2aa\") " pod="manila-kuttl-tests/manila-share-share1-0" Feb 02 15:38:07 crc kubenswrapper[4733]: I0202 15:38:07.117123 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f33ab3b4-cc56-4046-a8c5-bd859c2df2aa-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"f33ab3b4-cc56-4046-a8c5-bd859c2df2aa\") " pod="manila-kuttl-tests/manila-share-share1-0" Feb 02 15:38:07 crc kubenswrapper[4733]: I0202 15:38:07.117216 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f33ab3b4-cc56-4046-a8c5-bd859c2df2aa-ceph\") pod \"manila-share-share1-0\" (UID: \"f33ab3b4-cc56-4046-a8c5-bd859c2df2aa\") " pod="manila-kuttl-tests/manila-share-share1-0" Feb 02 15:38:07 crc kubenswrapper[4733]: I0202 15:38:07.117243 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/f33ab3b4-cc56-4046-a8c5-bd859c2df2aa-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"f33ab3b4-cc56-4046-a8c5-bd859c2df2aa\") " pod="manila-kuttl-tests/manila-share-share1-0" Feb 02 15:38:07 crc kubenswrapper[4733]: I0202 15:38:07.117267 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzm4m\" (UniqueName: \"kubernetes.io/projected/f33ab3b4-cc56-4046-a8c5-bd859c2df2aa-kube-api-access-tzm4m\") pod \"manila-share-share1-0\" (UID: \"f33ab3b4-cc56-4046-a8c5-bd859c2df2aa\") " pod="manila-kuttl-tests/manila-share-share1-0" Feb 02 15:38:07 crc kubenswrapper[4733]: I0202 15:38:07.117290 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f33ab3b4-cc56-4046-a8c5-bd859c2df2aa-config-data\") pod \"manila-share-share1-0\" (UID: \"f33ab3b4-cc56-4046-a8c5-bd859c2df2aa\") " pod="manila-kuttl-tests/manila-share-share1-0" Feb 02 15:38:07 crc kubenswrapper[4733]: I0202 15:38:07.117340 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f33ab3b4-cc56-4046-a8c5-bd859c2df2aa-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"f33ab3b4-cc56-4046-a8c5-bd859c2df2aa\") " pod="manila-kuttl-tests/manila-share-share1-0" Feb 02 15:38:07 crc kubenswrapper[4733]: I0202 15:38:07.218447 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f33ab3b4-cc56-4046-a8c5-bd859c2df2aa-scripts\") pod \"manila-share-share1-0\" (UID: \"f33ab3b4-cc56-4046-a8c5-bd859c2df2aa\") " pod="manila-kuttl-tests/manila-share-share1-0" Feb 02 15:38:07 crc kubenswrapper[4733]: I0202 15:38:07.218510 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f33ab3b4-cc56-4046-a8c5-bd859c2df2aa-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"f33ab3b4-cc56-4046-a8c5-bd859c2df2aa\") " pod="manila-kuttl-tests/manila-share-share1-0" Feb 02 15:38:07 crc kubenswrapper[4733]: I0202 15:38:07.218539 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f33ab3b4-cc56-4046-a8c5-bd859c2df2aa-ceph\") pod \"manila-share-share1-0\" (UID: \"f33ab3b4-cc56-4046-a8c5-bd859c2df2aa\") " pod="manila-kuttl-tests/manila-share-share1-0" Feb 02 15:38:07 crc kubenswrapper[4733]: I0202 15:38:07.218561 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/f33ab3b4-cc56-4046-a8c5-bd859c2df2aa-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"f33ab3b4-cc56-4046-a8c5-bd859c2df2aa\") " pod="manila-kuttl-tests/manila-share-share1-0" Feb 02 15:38:07 crc kubenswrapper[4733]: I0202 15:38:07.218581 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzm4m\" (UniqueName: \"kubernetes.io/projected/f33ab3b4-cc56-4046-a8c5-bd859c2df2aa-kube-api-access-tzm4m\") pod \"manila-share-share1-0\" (UID: \"f33ab3b4-cc56-4046-a8c5-bd859c2df2aa\") " pod="manila-kuttl-tests/manila-share-share1-0" Feb 02 15:38:07 crc kubenswrapper[4733]: I0202 15:38:07.218603 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f33ab3b4-cc56-4046-a8c5-bd859c2df2aa-config-data\") pod \"manila-share-share1-0\" (UID: \"f33ab3b4-cc56-4046-a8c5-bd859c2df2aa\") " pod="manila-kuttl-tests/manila-share-share1-0" Feb 02 15:38:07 crc kubenswrapper[4733]: I0202 15:38:07.218628 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f33ab3b4-cc56-4046-a8c5-bd859c2df2aa-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"f33ab3b4-cc56-4046-a8c5-bd859c2df2aa\") " pod="manila-kuttl-tests/manila-share-share1-0" Feb 02 15:38:07 crc kubenswrapper[4733]: I0202 15:38:07.218683 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f33ab3b4-cc56-4046-a8c5-bd859c2df2aa-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"f33ab3b4-cc56-4046-a8c5-bd859c2df2aa\") " pod="manila-kuttl-tests/manila-share-share1-0" Feb 02 15:38:07 crc kubenswrapper[4733]: I0202 15:38:07.218744 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/f33ab3b4-cc56-4046-a8c5-bd859c2df2aa-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"f33ab3b4-cc56-4046-a8c5-bd859c2df2aa\") " pod="manila-kuttl-tests/manila-share-share1-0" Feb 02 15:38:07 crc kubenswrapper[4733]: I0202 15:38:07.224343 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f33ab3b4-cc56-4046-a8c5-bd859c2df2aa-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"f33ab3b4-cc56-4046-a8c5-bd859c2df2aa\") " pod="manila-kuttl-tests/manila-share-share1-0" Feb 02 15:38:07 crc kubenswrapper[4733]: I0202 15:38:07.225313 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f33ab3b4-cc56-4046-a8c5-bd859c2df2aa-scripts\") pod \"manila-share-share1-0\" (UID: \"f33ab3b4-cc56-4046-a8c5-bd859c2df2aa\") " pod="manila-kuttl-tests/manila-share-share1-0" Feb 02 15:38:07 crc kubenswrapper[4733]: I0202 15:38:07.225493 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f33ab3b4-cc56-4046-a8c5-bd859c2df2aa-ceph\") pod \"manila-share-share1-0\" (UID: \"f33ab3b4-cc56-4046-a8c5-bd859c2df2aa\") " pod="manila-kuttl-tests/manila-share-share1-0" Feb 02 15:38:07 crc kubenswrapper[4733]: I0202 15:38:07.231502 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f33ab3b4-cc56-4046-a8c5-bd859c2df2aa-config-data\") pod \"manila-share-share1-0\" (UID: \"f33ab3b4-cc56-4046-a8c5-bd859c2df2aa\") " pod="manila-kuttl-tests/manila-share-share1-0" Feb 02 15:38:07 crc kubenswrapper[4733]: I0202 15:38:07.235670 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzm4m\" (UniqueName: \"kubernetes.io/projected/f33ab3b4-cc56-4046-a8c5-bd859c2df2aa-kube-api-access-tzm4m\") pod \"manila-share-share1-0\" (UID: \"f33ab3b4-cc56-4046-a8c5-bd859c2df2aa\") " pod="manila-kuttl-tests/manila-share-share1-0" Feb 02 15:38:07 crc kubenswrapper[4733]: I0202 15:38:07.351963 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-share-share1-0" Feb 02 15:38:07 crc kubenswrapper[4733]: I0202 15:38:07.900928 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/manila-share-share1-0"] Feb 02 15:38:08 crc kubenswrapper[4733]: I0202 15:38:08.266238 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-share-share1-0" event={"ID":"f33ab3b4-cc56-4046-a8c5-bd859c2df2aa","Type":"ContainerStarted","Data":"db5510313cce5956f185323bca26f48c0b1129157b40cd542257160d19914896"} Feb 02 15:38:09 crc kubenswrapper[4733]: I0202 15:38:09.283604 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-share-share1-0" event={"ID":"f33ab3b4-cc56-4046-a8c5-bd859c2df2aa","Type":"ContainerStarted","Data":"f2b8b4d6fef51440a511d27ecdf688ea096171d5dba203231c6cfffcb4880fd6"} Feb 02 15:38:09 crc kubenswrapper[4733]: I0202 15:38:09.283671 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-share-share1-0" event={"ID":"f33ab3b4-cc56-4046-a8c5-bd859c2df2aa","Type":"ContainerStarted","Data":"3423380ed77bc3c40a9c38ef88b85589d65ebcfa3856ed80eb5d3e2993383c89"} Feb 02 15:38:09 crc kubenswrapper[4733]: I0202 15:38:09.305218 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="manila-kuttl-tests/manila-share-share1-0" podStartSLOduration=2.305204205 podStartE2EDuration="2.305204205s" podCreationTimestamp="2026-02-02 15:38:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:38:09.302833983 +0000 UTC m=+1372.754295341" watchObservedRunningTime="2026-02-02 15:38:09.305204205 +0000 UTC m=+1372.756665563" Feb 02 15:38:12 crc kubenswrapper[4733]: I0202 15:38:12.066887 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zg4qj"] Feb 02 15:38:12 crc kubenswrapper[4733]: I0202 15:38:12.070209 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zg4qj" Feb 02 15:38:12 crc kubenswrapper[4733]: I0202 15:38:12.094474 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zg4qj"] Feb 02 15:38:12 crc kubenswrapper[4733]: I0202 15:38:12.194235 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77f7b81a-b74e-41a2-be1a-065e152e7df2-utilities\") pod \"redhat-operators-zg4qj\" (UID: \"77f7b81a-b74e-41a2-be1a-065e152e7df2\") " pod="openshift-marketplace/redhat-operators-zg4qj" Feb 02 15:38:12 crc kubenswrapper[4733]: I0202 15:38:12.194535 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqhg4\" (UniqueName: \"kubernetes.io/projected/77f7b81a-b74e-41a2-be1a-065e152e7df2-kube-api-access-sqhg4\") pod \"redhat-operators-zg4qj\" (UID: \"77f7b81a-b74e-41a2-be1a-065e152e7df2\") " pod="openshift-marketplace/redhat-operators-zg4qj" Feb 02 15:38:12 crc kubenswrapper[4733]: I0202 15:38:12.194836 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77f7b81a-b74e-41a2-be1a-065e152e7df2-catalog-content\") pod \"redhat-operators-zg4qj\" (UID: \"77f7b81a-b74e-41a2-be1a-065e152e7df2\") " pod="openshift-marketplace/redhat-operators-zg4qj" Feb 02 15:38:12 crc kubenswrapper[4733]: I0202 15:38:12.296710 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77f7b81a-b74e-41a2-be1a-065e152e7df2-catalog-content\") pod \"redhat-operators-zg4qj\" (UID: \"77f7b81a-b74e-41a2-be1a-065e152e7df2\") " pod="openshift-marketplace/redhat-operators-zg4qj" Feb 02 15:38:12 crc kubenswrapper[4733]: I0202 15:38:12.297532 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77f7b81a-b74e-41a2-be1a-065e152e7df2-catalog-content\") pod \"redhat-operators-zg4qj\" (UID: \"77f7b81a-b74e-41a2-be1a-065e152e7df2\") " pod="openshift-marketplace/redhat-operators-zg4qj" Feb 02 15:38:12 crc kubenswrapper[4733]: I0202 15:38:12.297627 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77f7b81a-b74e-41a2-be1a-065e152e7df2-utilities\") pod \"redhat-operators-zg4qj\" (UID: \"77f7b81a-b74e-41a2-be1a-065e152e7df2\") " pod="openshift-marketplace/redhat-operators-zg4qj" Feb 02 15:38:12 crc kubenswrapper[4733]: I0202 15:38:12.298024 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77f7b81a-b74e-41a2-be1a-065e152e7df2-utilities\") pod \"redhat-operators-zg4qj\" (UID: \"77f7b81a-b74e-41a2-be1a-065e152e7df2\") " pod="openshift-marketplace/redhat-operators-zg4qj" Feb 02 15:38:12 crc kubenswrapper[4733]: I0202 15:38:12.298546 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqhg4\" (UniqueName: \"kubernetes.io/projected/77f7b81a-b74e-41a2-be1a-065e152e7df2-kube-api-access-sqhg4\") pod \"redhat-operators-zg4qj\" (UID: \"77f7b81a-b74e-41a2-be1a-065e152e7df2\") " pod="openshift-marketplace/redhat-operators-zg4qj" Feb 02 15:38:12 crc kubenswrapper[4733]: I0202 15:38:12.323386 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqhg4\" (UniqueName: \"kubernetes.io/projected/77f7b81a-b74e-41a2-be1a-065e152e7df2-kube-api-access-sqhg4\") pod \"redhat-operators-zg4qj\" (UID: \"77f7b81a-b74e-41a2-be1a-065e152e7df2\") " pod="openshift-marketplace/redhat-operators-zg4qj" Feb 02 15:38:12 crc kubenswrapper[4733]: I0202 15:38:12.401102 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zg4qj" Feb 02 15:38:12 crc kubenswrapper[4733]: I0202 15:38:12.883482 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zg4qj"] Feb 02 15:38:12 crc kubenswrapper[4733]: W0202 15:38:12.887311 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod77f7b81a_b74e_41a2_be1a_065e152e7df2.slice/crio-1ec2ec1187db4a07b0468376c713004b0eeebf94dbe7c79eb8e4bcb7f84432e1 WatchSource:0}: Error finding container 1ec2ec1187db4a07b0468376c713004b0eeebf94dbe7c79eb8e4bcb7f84432e1: Status 404 returned error can't find the container with id 1ec2ec1187db4a07b0468376c713004b0eeebf94dbe7c79eb8e4bcb7f84432e1 Feb 02 15:38:13 crc kubenswrapper[4733]: I0202 15:38:13.321513 4733 generic.go:334] "Generic (PLEG): container finished" podID="77f7b81a-b74e-41a2-be1a-065e152e7df2" containerID="216e57641a8246bb773ef10d5a930b48a4e5c7f0a118ea242eee73529c30e36e" exitCode=0 Feb 02 15:38:13 crc kubenswrapper[4733]: I0202 15:38:13.321568 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zg4qj" event={"ID":"77f7b81a-b74e-41a2-be1a-065e152e7df2","Type":"ContainerDied","Data":"216e57641a8246bb773ef10d5a930b48a4e5c7f0a118ea242eee73529c30e36e"} Feb 02 15:38:13 crc kubenswrapper[4733]: I0202 15:38:13.321621 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zg4qj" event={"ID":"77f7b81a-b74e-41a2-be1a-065e152e7df2","Type":"ContainerStarted","Data":"1ec2ec1187db4a07b0468376c713004b0eeebf94dbe7c79eb8e4bcb7f84432e1"} Feb 02 15:38:13 crc kubenswrapper[4733]: I0202 15:38:13.325457 4733 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 02 15:38:15 crc kubenswrapper[4733]: I0202 15:38:15.341744 4733 generic.go:334] "Generic (PLEG): container finished" podID="77f7b81a-b74e-41a2-be1a-065e152e7df2" containerID="8fe25f86bec5e60c73d0bfaa6c98d148337d99c606b8fcaea3ee0776435e4aab" exitCode=0 Feb 02 15:38:15 crc kubenswrapper[4733]: I0202 15:38:15.341826 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zg4qj" event={"ID":"77f7b81a-b74e-41a2-be1a-065e152e7df2","Type":"ContainerDied","Data":"8fe25f86bec5e60c73d0bfaa6c98d148337d99c606b8fcaea3ee0776435e4aab"} Feb 02 15:38:16 crc kubenswrapper[4733]: I0202 15:38:16.352312 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zg4qj" event={"ID":"77f7b81a-b74e-41a2-be1a-065e152e7df2","Type":"ContainerStarted","Data":"024ce8576a4037de81134b65c71f1aef864d2bd0300f10e77d3375c981e4dec2"} Feb 02 15:38:16 crc kubenswrapper[4733]: I0202 15:38:16.379625 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zg4qj" podStartSLOduration=1.918810291 podStartE2EDuration="4.379610115s" podCreationTimestamp="2026-02-02 15:38:12 +0000 UTC" firstStartedPulling="2026-02-02 15:38:13.325077453 +0000 UTC m=+1376.776538821" lastFinishedPulling="2026-02-02 15:38:15.785877247 +0000 UTC m=+1379.237338645" observedRunningTime="2026-02-02 15:38:16.375132579 +0000 UTC m=+1379.826593947" watchObservedRunningTime="2026-02-02 15:38:16.379610115 +0000 UTC m=+1379.831071473" Feb 02 15:38:17 crc kubenswrapper[4733]: I0202 15:38:17.352972 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="manila-kuttl-tests/manila-share-share1-0" Feb 02 15:38:22 crc kubenswrapper[4733]: I0202 15:38:22.402324 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zg4qj" Feb 02 15:38:22 crc kubenswrapper[4733]: I0202 15:38:22.402812 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zg4qj" Feb 02 15:38:23 crc kubenswrapper[4733]: I0202 15:38:23.453104 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-zg4qj" podUID="77f7b81a-b74e-41a2-be1a-065e152e7df2" containerName="registry-server" probeResult="failure" output=< Feb 02 15:38:23 crc kubenswrapper[4733]: timeout: failed to connect service ":50051" within 1s Feb 02 15:38:23 crc kubenswrapper[4733]: > Feb 02 15:38:28 crc kubenswrapper[4733]: I0202 15:38:28.787389 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="manila-kuttl-tests/manila-share-share1-0" Feb 02 15:38:30 crc kubenswrapper[4733]: I0202 15:38:30.090413 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/manila-share-share0-0"] Feb 02 15:38:30 crc kubenswrapper[4733]: I0202 15:38:30.090643 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="manila-kuttl-tests/manila-share-share0-0" podUID="2267a854-0257-4ab1-8951-b82dc2b7350f" containerName="manila-share" containerID="cri-o://8127c0a93057ef0939fd26da8e2f58eadf4e6ae1f51f0bfb3942c07ad85a5eed" gracePeriod=30 Feb 02 15:38:30 crc kubenswrapper[4733]: I0202 15:38:30.090975 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="manila-kuttl-tests/manila-share-share0-0" podUID="2267a854-0257-4ab1-8951-b82dc2b7350f" containerName="probe" containerID="cri-o://8be19e6ccae7070cbdb2e63b741acd61dd3e9c876427ca89a3f9e49547ab4db1" gracePeriod=30 Feb 02 15:38:30 crc kubenswrapper[4733]: I0202 15:38:30.465126 4733 generic.go:334] "Generic (PLEG): container finished" podID="2267a854-0257-4ab1-8951-b82dc2b7350f" containerID="8be19e6ccae7070cbdb2e63b741acd61dd3e9c876427ca89a3f9e49547ab4db1" exitCode=0 Feb 02 15:38:30 crc kubenswrapper[4733]: I0202 15:38:30.465219 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-share-share0-0" event={"ID":"2267a854-0257-4ab1-8951-b82dc2b7350f","Type":"ContainerDied","Data":"8be19e6ccae7070cbdb2e63b741acd61dd3e9c876427ca89a3f9e49547ab4db1"} Feb 02 15:38:30 crc kubenswrapper[4733]: I0202 15:38:30.936094 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:38:31 crc kubenswrapper[4733]: I0202 15:38:31.010924 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/2267a854-0257-4ab1-8951-b82dc2b7350f-var-lib-manila\") pod \"2267a854-0257-4ab1-8951-b82dc2b7350f\" (UID: \"2267a854-0257-4ab1-8951-b82dc2b7350f\") " Feb 02 15:38:31 crc kubenswrapper[4733]: I0202 15:38:31.010984 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2267a854-0257-4ab1-8951-b82dc2b7350f-ceph\") pod \"2267a854-0257-4ab1-8951-b82dc2b7350f\" (UID: \"2267a854-0257-4ab1-8951-b82dc2b7350f\") " Feb 02 15:38:31 crc kubenswrapper[4733]: I0202 15:38:31.011007 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2267a854-0257-4ab1-8951-b82dc2b7350f-config-data-custom\") pod \"2267a854-0257-4ab1-8951-b82dc2b7350f\" (UID: \"2267a854-0257-4ab1-8951-b82dc2b7350f\") " Feb 02 15:38:31 crc kubenswrapper[4733]: I0202 15:38:31.011041 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2267a854-0257-4ab1-8951-b82dc2b7350f-scripts\") pod \"2267a854-0257-4ab1-8951-b82dc2b7350f\" (UID: \"2267a854-0257-4ab1-8951-b82dc2b7350f\") " Feb 02 15:38:31 crc kubenswrapper[4733]: I0202 15:38:31.011074 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-52l96\" (UniqueName: \"kubernetes.io/projected/2267a854-0257-4ab1-8951-b82dc2b7350f-kube-api-access-52l96\") pod \"2267a854-0257-4ab1-8951-b82dc2b7350f\" (UID: \"2267a854-0257-4ab1-8951-b82dc2b7350f\") " Feb 02 15:38:31 crc kubenswrapper[4733]: I0202 15:38:31.011135 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2267a854-0257-4ab1-8951-b82dc2b7350f-config-data\") pod \"2267a854-0257-4ab1-8951-b82dc2b7350f\" (UID: \"2267a854-0257-4ab1-8951-b82dc2b7350f\") " Feb 02 15:38:31 crc kubenswrapper[4733]: I0202 15:38:31.011203 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2267a854-0257-4ab1-8951-b82dc2b7350f-etc-machine-id\") pod \"2267a854-0257-4ab1-8951-b82dc2b7350f\" (UID: \"2267a854-0257-4ab1-8951-b82dc2b7350f\") " Feb 02 15:38:31 crc kubenswrapper[4733]: I0202 15:38:31.011477 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2267a854-0257-4ab1-8951-b82dc2b7350f-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "2267a854-0257-4ab1-8951-b82dc2b7350f" (UID: "2267a854-0257-4ab1-8951-b82dc2b7350f"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 15:38:31 crc kubenswrapper[4733]: I0202 15:38:31.011512 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2267a854-0257-4ab1-8951-b82dc2b7350f-var-lib-manila" (OuterVolumeSpecName: "var-lib-manila") pod "2267a854-0257-4ab1-8951-b82dc2b7350f" (UID: "2267a854-0257-4ab1-8951-b82dc2b7350f"). InnerVolumeSpecName "var-lib-manila". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 15:38:31 crc kubenswrapper[4733]: I0202 15:38:31.018021 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2267a854-0257-4ab1-8951-b82dc2b7350f-ceph" (OuterVolumeSpecName: "ceph") pod "2267a854-0257-4ab1-8951-b82dc2b7350f" (UID: "2267a854-0257-4ab1-8951-b82dc2b7350f"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:38:31 crc kubenswrapper[4733]: I0202 15:38:31.018047 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2267a854-0257-4ab1-8951-b82dc2b7350f-kube-api-access-52l96" (OuterVolumeSpecName: "kube-api-access-52l96") pod "2267a854-0257-4ab1-8951-b82dc2b7350f" (UID: "2267a854-0257-4ab1-8951-b82dc2b7350f"). InnerVolumeSpecName "kube-api-access-52l96". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:38:31 crc kubenswrapper[4733]: I0202 15:38:31.019288 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2267a854-0257-4ab1-8951-b82dc2b7350f-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "2267a854-0257-4ab1-8951-b82dc2b7350f" (UID: "2267a854-0257-4ab1-8951-b82dc2b7350f"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:38:31 crc kubenswrapper[4733]: I0202 15:38:31.020005 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2267a854-0257-4ab1-8951-b82dc2b7350f-scripts" (OuterVolumeSpecName: "scripts") pod "2267a854-0257-4ab1-8951-b82dc2b7350f" (UID: "2267a854-0257-4ab1-8951-b82dc2b7350f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:38:31 crc kubenswrapper[4733]: I0202 15:38:31.083481 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2267a854-0257-4ab1-8951-b82dc2b7350f-config-data" (OuterVolumeSpecName: "config-data") pod "2267a854-0257-4ab1-8951-b82dc2b7350f" (UID: "2267a854-0257-4ab1-8951-b82dc2b7350f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:38:31 crc kubenswrapper[4733]: I0202 15:38:31.112607 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2267a854-0257-4ab1-8951-b82dc2b7350f-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:31 crc kubenswrapper[4733]: I0202 15:38:31.112722 4733 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2267a854-0257-4ab1-8951-b82dc2b7350f-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:31 crc kubenswrapper[4733]: I0202 15:38:31.112748 4733 reconciler_common.go:293] "Volume detached for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/2267a854-0257-4ab1-8951-b82dc2b7350f-var-lib-manila\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:31 crc kubenswrapper[4733]: I0202 15:38:31.112760 4733 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2267a854-0257-4ab1-8951-b82dc2b7350f-ceph\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:31 crc kubenswrapper[4733]: I0202 15:38:31.112774 4733 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2267a854-0257-4ab1-8951-b82dc2b7350f-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:31 crc kubenswrapper[4733]: I0202 15:38:31.112786 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2267a854-0257-4ab1-8951-b82dc2b7350f-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:31 crc kubenswrapper[4733]: I0202 15:38:31.112797 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-52l96\" (UniqueName: \"kubernetes.io/projected/2267a854-0257-4ab1-8951-b82dc2b7350f-kube-api-access-52l96\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:31 crc kubenswrapper[4733]: I0202 15:38:31.478316 4733 generic.go:334] "Generic (PLEG): container finished" podID="2267a854-0257-4ab1-8951-b82dc2b7350f" containerID="8127c0a93057ef0939fd26da8e2f58eadf4e6ae1f51f0bfb3942c07ad85a5eed" exitCode=1 Feb 02 15:38:31 crc kubenswrapper[4733]: I0202 15:38:31.478410 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-share-share0-0" Feb 02 15:38:31 crc kubenswrapper[4733]: I0202 15:38:31.478414 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-share-share0-0" event={"ID":"2267a854-0257-4ab1-8951-b82dc2b7350f","Type":"ContainerDied","Data":"8127c0a93057ef0939fd26da8e2f58eadf4e6ae1f51f0bfb3942c07ad85a5eed"} Feb 02 15:38:31 crc kubenswrapper[4733]: I0202 15:38:31.479254 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-share-share0-0" event={"ID":"2267a854-0257-4ab1-8951-b82dc2b7350f","Type":"ContainerDied","Data":"289c652f6d515b9cbad9ecb79e7fa768638c8017c5229ce454d853c166fc18a0"} Feb 02 15:38:31 crc kubenswrapper[4733]: I0202 15:38:31.479297 4733 scope.go:117] "RemoveContainer" containerID="8be19e6ccae7070cbdb2e63b741acd61dd3e9c876427ca89a3f9e49547ab4db1" Feb 02 15:38:31 crc kubenswrapper[4733]: I0202 15:38:31.508324 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/manila-share-share0-0"] Feb 02 15:38:31 crc kubenswrapper[4733]: I0202 15:38:31.512842 4733 scope.go:117] "RemoveContainer" containerID="8127c0a93057ef0939fd26da8e2f58eadf4e6ae1f51f0bfb3942c07ad85a5eed" Feb 02 15:38:31 crc kubenswrapper[4733]: I0202 15:38:31.518744 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["manila-kuttl-tests/manila-share-share0-0"] Feb 02 15:38:31 crc kubenswrapper[4733]: I0202 15:38:31.538519 4733 scope.go:117] "RemoveContainer" containerID="8be19e6ccae7070cbdb2e63b741acd61dd3e9c876427ca89a3f9e49547ab4db1" Feb 02 15:38:31 crc kubenswrapper[4733]: E0202 15:38:31.539352 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8be19e6ccae7070cbdb2e63b741acd61dd3e9c876427ca89a3f9e49547ab4db1\": container with ID starting with 8be19e6ccae7070cbdb2e63b741acd61dd3e9c876427ca89a3f9e49547ab4db1 not found: ID does not exist" containerID="8be19e6ccae7070cbdb2e63b741acd61dd3e9c876427ca89a3f9e49547ab4db1" Feb 02 15:38:31 crc kubenswrapper[4733]: I0202 15:38:31.539410 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8be19e6ccae7070cbdb2e63b741acd61dd3e9c876427ca89a3f9e49547ab4db1"} err="failed to get container status \"8be19e6ccae7070cbdb2e63b741acd61dd3e9c876427ca89a3f9e49547ab4db1\": rpc error: code = NotFound desc = could not find container \"8be19e6ccae7070cbdb2e63b741acd61dd3e9c876427ca89a3f9e49547ab4db1\": container with ID starting with 8be19e6ccae7070cbdb2e63b741acd61dd3e9c876427ca89a3f9e49547ab4db1 not found: ID does not exist" Feb 02 15:38:31 crc kubenswrapper[4733]: I0202 15:38:31.539441 4733 scope.go:117] "RemoveContainer" containerID="8127c0a93057ef0939fd26da8e2f58eadf4e6ae1f51f0bfb3942c07ad85a5eed" Feb 02 15:38:31 crc kubenswrapper[4733]: E0202 15:38:31.540948 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8127c0a93057ef0939fd26da8e2f58eadf4e6ae1f51f0bfb3942c07ad85a5eed\": container with ID starting with 8127c0a93057ef0939fd26da8e2f58eadf4e6ae1f51f0bfb3942c07ad85a5eed not found: ID does not exist" containerID="8127c0a93057ef0939fd26da8e2f58eadf4e6ae1f51f0bfb3942c07ad85a5eed" Feb 02 15:38:31 crc kubenswrapper[4733]: I0202 15:38:31.541031 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8127c0a93057ef0939fd26da8e2f58eadf4e6ae1f51f0bfb3942c07ad85a5eed"} err="failed to get container status \"8127c0a93057ef0939fd26da8e2f58eadf4e6ae1f51f0bfb3942c07ad85a5eed\": rpc error: code = NotFound desc = could not find container \"8127c0a93057ef0939fd26da8e2f58eadf4e6ae1f51f0bfb3942c07ad85a5eed\": container with ID starting with 8127c0a93057ef0939fd26da8e2f58eadf4e6ae1f51f0bfb3942c07ad85a5eed not found: ID does not exist" Feb 02 15:38:32 crc kubenswrapper[4733]: I0202 15:38:32.472679 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zg4qj" Feb 02 15:38:32 crc kubenswrapper[4733]: I0202 15:38:32.532438 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["manila-kuttl-tests/manila-service-cleanup-n5b5h655-8d2b7"] Feb 02 15:38:32 crc kubenswrapper[4733]: E0202 15:38:32.532846 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2267a854-0257-4ab1-8951-b82dc2b7350f" containerName="probe" Feb 02 15:38:32 crc kubenswrapper[4733]: I0202 15:38:32.532879 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="2267a854-0257-4ab1-8951-b82dc2b7350f" containerName="probe" Feb 02 15:38:32 crc kubenswrapper[4733]: E0202 15:38:32.532922 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2267a854-0257-4ab1-8951-b82dc2b7350f" containerName="manila-share" Feb 02 15:38:32 crc kubenswrapper[4733]: I0202 15:38:32.532934 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="2267a854-0257-4ab1-8951-b82dc2b7350f" containerName="manila-share" Feb 02 15:38:32 crc kubenswrapper[4733]: I0202 15:38:32.533132 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="2267a854-0257-4ab1-8951-b82dc2b7350f" containerName="probe" Feb 02 15:38:32 crc kubenswrapper[4733]: I0202 15:38:32.533195 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="2267a854-0257-4ab1-8951-b82dc2b7350f" containerName="manila-share" Feb 02 15:38:32 crc kubenswrapper[4733]: I0202 15:38:32.533917 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-service-cleanup-n5b5h655-8d2b7" Feb 02 15:38:32 crc kubenswrapper[4733]: I0202 15:38:32.557394 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/manila-service-cleanup-n5b5h655-8d2b7"] Feb 02 15:38:32 crc kubenswrapper[4733]: I0202 15:38:32.557470 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zg4qj" Feb 02 15:38:32 crc kubenswrapper[4733]: I0202 15:38:32.632011 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9dffed5a-a57d-4710-909f-799b47ff68ef-config-data\") pod \"manila-service-cleanup-n5b5h655-8d2b7\" (UID: \"9dffed5a-a57d-4710-909f-799b47ff68ef\") " pod="manila-kuttl-tests/manila-service-cleanup-n5b5h655-8d2b7" Feb 02 15:38:32 crc kubenswrapper[4733]: I0202 15:38:32.632105 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lms2\" (UniqueName: \"kubernetes.io/projected/9dffed5a-a57d-4710-909f-799b47ff68ef-kube-api-access-8lms2\") pod \"manila-service-cleanup-n5b5h655-8d2b7\" (UID: \"9dffed5a-a57d-4710-909f-799b47ff68ef\") " pod="manila-kuttl-tests/manila-service-cleanup-n5b5h655-8d2b7" Feb 02 15:38:32 crc kubenswrapper[4733]: I0202 15:38:32.632154 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/9dffed5a-a57d-4710-909f-799b47ff68ef-job-config-data\") pod \"manila-service-cleanup-n5b5h655-8d2b7\" (UID: \"9dffed5a-a57d-4710-909f-799b47ff68ef\") " pod="manila-kuttl-tests/manila-service-cleanup-n5b5h655-8d2b7" Feb 02 15:38:32 crc kubenswrapper[4733]: I0202 15:38:32.734196 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9dffed5a-a57d-4710-909f-799b47ff68ef-config-data\") pod \"manila-service-cleanup-n5b5h655-8d2b7\" (UID: \"9dffed5a-a57d-4710-909f-799b47ff68ef\") " pod="manila-kuttl-tests/manila-service-cleanup-n5b5h655-8d2b7" Feb 02 15:38:32 crc kubenswrapper[4733]: I0202 15:38:32.734279 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lms2\" (UniqueName: \"kubernetes.io/projected/9dffed5a-a57d-4710-909f-799b47ff68ef-kube-api-access-8lms2\") pod \"manila-service-cleanup-n5b5h655-8d2b7\" (UID: \"9dffed5a-a57d-4710-909f-799b47ff68ef\") " pod="manila-kuttl-tests/manila-service-cleanup-n5b5h655-8d2b7" Feb 02 15:38:32 crc kubenswrapper[4733]: I0202 15:38:32.734321 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/9dffed5a-a57d-4710-909f-799b47ff68ef-job-config-data\") pod \"manila-service-cleanup-n5b5h655-8d2b7\" (UID: \"9dffed5a-a57d-4710-909f-799b47ff68ef\") " pod="manila-kuttl-tests/manila-service-cleanup-n5b5h655-8d2b7" Feb 02 15:38:32 crc kubenswrapper[4733]: I0202 15:38:32.735087 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zg4qj"] Feb 02 15:38:32 crc kubenswrapper[4733]: I0202 15:38:32.740410 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9dffed5a-a57d-4710-909f-799b47ff68ef-config-data\") pod \"manila-service-cleanup-n5b5h655-8d2b7\" (UID: \"9dffed5a-a57d-4710-909f-799b47ff68ef\") " pod="manila-kuttl-tests/manila-service-cleanup-n5b5h655-8d2b7" Feb 02 15:38:32 crc kubenswrapper[4733]: I0202 15:38:32.745969 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/9dffed5a-a57d-4710-909f-799b47ff68ef-job-config-data\") pod \"manila-service-cleanup-n5b5h655-8d2b7\" (UID: \"9dffed5a-a57d-4710-909f-799b47ff68ef\") " pod="manila-kuttl-tests/manila-service-cleanup-n5b5h655-8d2b7" Feb 02 15:38:32 crc kubenswrapper[4733]: I0202 15:38:32.760301 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lms2\" (UniqueName: \"kubernetes.io/projected/9dffed5a-a57d-4710-909f-799b47ff68ef-kube-api-access-8lms2\") pod \"manila-service-cleanup-n5b5h655-8d2b7\" (UID: \"9dffed5a-a57d-4710-909f-799b47ff68ef\") " pod="manila-kuttl-tests/manila-service-cleanup-n5b5h655-8d2b7" Feb 02 15:38:32 crc kubenswrapper[4733]: I0202 15:38:32.868187 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-service-cleanup-n5b5h655-8d2b7" Feb 02 15:38:33 crc kubenswrapper[4733]: I0202 15:38:33.061254 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/manila-db-sync-8q2p6"] Feb 02 15:38:33 crc kubenswrapper[4733]: I0202 15:38:33.068346 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/manila-service-cleanup-n5b5h655-8d2b7"] Feb 02 15:38:33 crc kubenswrapper[4733]: I0202 15:38:33.073639 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["manila-kuttl-tests/manila-db-sync-8q2p6"] Feb 02 15:38:33 crc kubenswrapper[4733]: I0202 15:38:33.119381 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/manila-share-share1-0"] Feb 02 15:38:33 crc kubenswrapper[4733]: I0202 15:38:33.119632 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="manila-kuttl-tests/manila-share-share1-0" podUID="f33ab3b4-cc56-4046-a8c5-bd859c2df2aa" containerName="manila-share" containerID="cri-o://3423380ed77bc3c40a9c38ef88b85589d65ebcfa3856ed80eb5d3e2993383c89" gracePeriod=30 Feb 02 15:38:33 crc kubenswrapper[4733]: I0202 15:38:33.119781 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="manila-kuttl-tests/manila-share-share1-0" podUID="f33ab3b4-cc56-4046-a8c5-bd859c2df2aa" containerName="probe" containerID="cri-o://f2b8b4d6fef51440a511d27ecdf688ea096171d5dba203231c6cfffcb4880fd6" gracePeriod=30 Feb 02 15:38:33 crc kubenswrapper[4733]: I0202 15:38:33.136584 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/manila-scheduler-0"] Feb 02 15:38:33 crc kubenswrapper[4733]: I0202 15:38:33.136811 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="manila-kuttl-tests/manila-scheduler-0" podUID="9abf907e-85df-4616-a830-56fae1ae3f69" containerName="manila-scheduler" containerID="cri-o://1b4db59339def5c0775f38a97cbcb2c76ad7e0dde03ca35dd3b89b993012719d" gracePeriod=30 Feb 02 15:38:33 crc kubenswrapper[4733]: I0202 15:38:33.137145 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="manila-kuttl-tests/manila-scheduler-0" podUID="9abf907e-85df-4616-a830-56fae1ae3f69" containerName="probe" containerID="cri-o://b7e17883ce2314948330a91af64d17c2235f67db1c6a7c62d5c1097a7e2be57f" gracePeriod=30 Feb 02 15:38:33 crc kubenswrapper[4733]: I0202 15:38:33.145619 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["manila-kuttl-tests/manilaa9fc-account-delete-fb8w5"] Feb 02 15:38:33 crc kubenswrapper[4733]: I0202 15:38:33.146788 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manilaa9fc-account-delete-fb8w5" Feb 02 15:38:33 crc kubenswrapper[4733]: I0202 15:38:33.162024 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/manilaa9fc-account-delete-fb8w5"] Feb 02 15:38:33 crc kubenswrapper[4733]: I0202 15:38:33.177379 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/manila-api-0"] Feb 02 15:38:33 crc kubenswrapper[4733]: I0202 15:38:33.177640 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="manila-kuttl-tests/manila-api-0" podUID="fe12e7c9-118b-4413-8fa4-96745802310e" containerName="manila-api-log" containerID="cri-o://11acc8fca457138b1b74f4ee6a9396a9e163b6a49c60cad0a5e39487d7377acf" gracePeriod=30 Feb 02 15:38:33 crc kubenswrapper[4733]: I0202 15:38:33.177680 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="manila-kuttl-tests/manila-api-0" podUID="fe12e7c9-118b-4413-8fa4-96745802310e" containerName="manila-api" containerID="cri-o://0c0517983e8f69418d96a083e721d09d7185d0fbb013e0b4fc4e93c6e1b2a795" gracePeriod=30 Feb 02 15:38:33 crc kubenswrapper[4733]: I0202 15:38:33.240714 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/62b4bf32-8cef-4a0c-bec7-c3511e5f8929-operator-scripts\") pod \"manilaa9fc-account-delete-fb8w5\" (UID: \"62b4bf32-8cef-4a0c-bec7-c3511e5f8929\") " pod="manila-kuttl-tests/manilaa9fc-account-delete-fb8w5" Feb 02 15:38:33 crc kubenswrapper[4733]: I0202 15:38:33.240843 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppl9m\" (UniqueName: \"kubernetes.io/projected/62b4bf32-8cef-4a0c-bec7-c3511e5f8929-kube-api-access-ppl9m\") pod \"manilaa9fc-account-delete-fb8w5\" (UID: \"62b4bf32-8cef-4a0c-bec7-c3511e5f8929\") " pod="manila-kuttl-tests/manilaa9fc-account-delete-fb8w5" Feb 02 15:38:33 crc kubenswrapper[4733]: I0202 15:38:33.267019 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e42c981-8832-49a6-b150-811064d2509b" path="/var/lib/kubelet/pods/0e42c981-8832-49a6-b150-811064d2509b/volumes" Feb 02 15:38:33 crc kubenswrapper[4733]: I0202 15:38:33.268049 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2267a854-0257-4ab1-8951-b82dc2b7350f" path="/var/lib/kubelet/pods/2267a854-0257-4ab1-8951-b82dc2b7350f/volumes" Feb 02 15:38:33 crc kubenswrapper[4733]: I0202 15:38:33.269779 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/manila-service-cleanup-n5b5h655-8d2b7"] Feb 02 15:38:33 crc kubenswrapper[4733]: I0202 15:38:33.342508 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/62b4bf32-8cef-4a0c-bec7-c3511e5f8929-operator-scripts\") pod \"manilaa9fc-account-delete-fb8w5\" (UID: \"62b4bf32-8cef-4a0c-bec7-c3511e5f8929\") " pod="manila-kuttl-tests/manilaa9fc-account-delete-fb8w5" Feb 02 15:38:33 crc kubenswrapper[4733]: I0202 15:38:33.342610 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppl9m\" (UniqueName: \"kubernetes.io/projected/62b4bf32-8cef-4a0c-bec7-c3511e5f8929-kube-api-access-ppl9m\") pod \"manilaa9fc-account-delete-fb8w5\" (UID: \"62b4bf32-8cef-4a0c-bec7-c3511e5f8929\") " pod="manila-kuttl-tests/manilaa9fc-account-delete-fb8w5" Feb 02 15:38:33 crc kubenswrapper[4733]: I0202 15:38:33.343809 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/62b4bf32-8cef-4a0c-bec7-c3511e5f8929-operator-scripts\") pod \"manilaa9fc-account-delete-fb8w5\" (UID: \"62b4bf32-8cef-4a0c-bec7-c3511e5f8929\") " pod="manila-kuttl-tests/manilaa9fc-account-delete-fb8w5" Feb 02 15:38:33 crc kubenswrapper[4733]: I0202 15:38:33.362888 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppl9m\" (UniqueName: \"kubernetes.io/projected/62b4bf32-8cef-4a0c-bec7-c3511e5f8929-kube-api-access-ppl9m\") pod \"manilaa9fc-account-delete-fb8w5\" (UID: \"62b4bf32-8cef-4a0c-bec7-c3511e5f8929\") " pod="manila-kuttl-tests/manilaa9fc-account-delete-fb8w5" Feb 02 15:38:33 crc kubenswrapper[4733]: I0202 15:38:33.479330 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manilaa9fc-account-delete-fb8w5" Feb 02 15:38:33 crc kubenswrapper[4733]: I0202 15:38:33.532708 4733 generic.go:334] "Generic (PLEG): container finished" podID="fe12e7c9-118b-4413-8fa4-96745802310e" containerID="11acc8fca457138b1b74f4ee6a9396a9e163b6a49c60cad0a5e39487d7377acf" exitCode=143 Feb 02 15:38:33 crc kubenswrapper[4733]: I0202 15:38:33.532805 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-api-0" event={"ID":"fe12e7c9-118b-4413-8fa4-96745802310e","Type":"ContainerDied","Data":"11acc8fca457138b1b74f4ee6a9396a9e163b6a49c60cad0a5e39487d7377acf"} Feb 02 15:38:33 crc kubenswrapper[4733]: I0202 15:38:33.535227 4733 generic.go:334] "Generic (PLEG): container finished" podID="f33ab3b4-cc56-4046-a8c5-bd859c2df2aa" containerID="f2b8b4d6fef51440a511d27ecdf688ea096171d5dba203231c6cfffcb4880fd6" exitCode=0 Feb 02 15:38:33 crc kubenswrapper[4733]: I0202 15:38:33.535284 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-share-share1-0" event={"ID":"f33ab3b4-cc56-4046-a8c5-bd859c2df2aa","Type":"ContainerDied","Data":"f2b8b4d6fef51440a511d27ecdf688ea096171d5dba203231c6cfffcb4880fd6"} Feb 02 15:38:33 crc kubenswrapper[4733]: I0202 15:38:33.540434 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-service-cleanup-n5b5h655-8d2b7" event={"ID":"9dffed5a-a57d-4710-909f-799b47ff68ef","Type":"ContainerStarted","Data":"75d93405a5f061a3a7692dc88c34cbd68d03aafedc47ed7489f01c130550f0ec"} Feb 02 15:38:33 crc kubenswrapper[4733]: I0202 15:38:33.540609 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-zg4qj" podUID="77f7b81a-b74e-41a2-be1a-065e152e7df2" containerName="registry-server" containerID="cri-o://024ce8576a4037de81134b65c71f1aef864d2bd0300f10e77d3375c981e4dec2" gracePeriod=2 Feb 02 15:38:33 crc kubenswrapper[4733]: I0202 15:38:33.691272 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/manilaa9fc-account-delete-fb8w5"] Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.020919 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/manila-operator-controller-manager-64f9574466-56vg6"] Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.021353 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/manila-operator-controller-manager-64f9574466-56vg6" podUID="f3971f39-e65f-4262-acaf-ca5fca3f7cc9" containerName="manager" containerID="cri-o://84e34713078767c3675090908ff02de8038baca92063e381c481b5e26d015c5b" gracePeriod=10 Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.112647 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-share-share1-0" Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.135617 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zg4qj" Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.178013 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tzm4m\" (UniqueName: \"kubernetes.io/projected/f33ab3b4-cc56-4046-a8c5-bd859c2df2aa-kube-api-access-tzm4m\") pod \"f33ab3b4-cc56-4046-a8c5-bd859c2df2aa\" (UID: \"f33ab3b4-cc56-4046-a8c5-bd859c2df2aa\") " Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.178106 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f33ab3b4-cc56-4046-a8c5-bd859c2df2aa-etc-machine-id\") pod \"f33ab3b4-cc56-4046-a8c5-bd859c2df2aa\" (UID: \"f33ab3b4-cc56-4046-a8c5-bd859c2df2aa\") " Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.178220 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f33ab3b4-cc56-4046-a8c5-bd859c2df2aa-scripts\") pod \"f33ab3b4-cc56-4046-a8c5-bd859c2df2aa\" (UID: \"f33ab3b4-cc56-4046-a8c5-bd859c2df2aa\") " Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.178270 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f33ab3b4-cc56-4046-a8c5-bd859c2df2aa-config-data-custom\") pod \"f33ab3b4-cc56-4046-a8c5-bd859c2df2aa\" (UID: \"f33ab3b4-cc56-4046-a8c5-bd859c2df2aa\") " Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.178378 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f33ab3b4-cc56-4046-a8c5-bd859c2df2aa-ceph\") pod \"f33ab3b4-cc56-4046-a8c5-bd859c2df2aa\" (UID: \"f33ab3b4-cc56-4046-a8c5-bd859c2df2aa\") " Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.178459 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/f33ab3b4-cc56-4046-a8c5-bd859c2df2aa-var-lib-manila\") pod \"f33ab3b4-cc56-4046-a8c5-bd859c2df2aa\" (UID: \"f33ab3b4-cc56-4046-a8c5-bd859c2df2aa\") " Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.178486 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f33ab3b4-cc56-4046-a8c5-bd859c2df2aa-config-data\") pod \"f33ab3b4-cc56-4046-a8c5-bd859c2df2aa\" (UID: \"f33ab3b4-cc56-4046-a8c5-bd859c2df2aa\") " Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.190880 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f33ab3b4-cc56-4046-a8c5-bd859c2df2aa-kube-api-access-tzm4m" (OuterVolumeSpecName: "kube-api-access-tzm4m") pod "f33ab3b4-cc56-4046-a8c5-bd859c2df2aa" (UID: "f33ab3b4-cc56-4046-a8c5-bd859c2df2aa"). InnerVolumeSpecName "kube-api-access-tzm4m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.191117 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f33ab3b4-cc56-4046-a8c5-bd859c2df2aa-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "f33ab3b4-cc56-4046-a8c5-bd859c2df2aa" (UID: "f33ab3b4-cc56-4046-a8c5-bd859c2df2aa"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.191258 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f33ab3b4-cc56-4046-a8c5-bd859c2df2aa-var-lib-manila" (OuterVolumeSpecName: "var-lib-manila") pod "f33ab3b4-cc56-4046-a8c5-bd859c2df2aa" (UID: "f33ab3b4-cc56-4046-a8c5-bd859c2df2aa"). InnerVolumeSpecName "var-lib-manila". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.217223 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f33ab3b4-cc56-4046-a8c5-bd859c2df2aa-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "f33ab3b4-cc56-4046-a8c5-bd859c2df2aa" (UID: "f33ab3b4-cc56-4046-a8c5-bd859c2df2aa"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.219252 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f33ab3b4-cc56-4046-a8c5-bd859c2df2aa-scripts" (OuterVolumeSpecName: "scripts") pod "f33ab3b4-cc56-4046-a8c5-bd859c2df2aa" (UID: "f33ab3b4-cc56-4046-a8c5-bd859c2df2aa"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.221445 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/manila-operator-index-gqz7t"] Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.221765 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/manila-operator-index-gqz7t" podUID="1ae26a29-79a7-4f0a-b546-f20743ba97a3" containerName="registry-server" containerID="cri-o://6b98411c753ef28c5622e28730c655cc9b2d9663f5b9c1e89989944c008b5759" gracePeriod=30 Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.224004 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f33ab3b4-cc56-4046-a8c5-bd859c2df2aa-ceph" (OuterVolumeSpecName: "ceph") pod "f33ab3b4-cc56-4046-a8c5-bd859c2df2aa" (UID: "f33ab3b4-cc56-4046-a8c5-bd859c2df2aa"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.257959 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/28815fdbd8da3955636af5641949aa0dc9d10a5ebc7db4178a6eac7a27cmqpd"] Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.264420 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/28815fdbd8da3955636af5641949aa0dc9d10a5ebc7db4178a6eac7a27cmqpd"] Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.291430 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77f7b81a-b74e-41a2-be1a-065e152e7df2-catalog-content\") pod \"77f7b81a-b74e-41a2-be1a-065e152e7df2\" (UID: \"77f7b81a-b74e-41a2-be1a-065e152e7df2\") " Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.291511 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77f7b81a-b74e-41a2-be1a-065e152e7df2-utilities\") pod \"77f7b81a-b74e-41a2-be1a-065e152e7df2\" (UID: \"77f7b81a-b74e-41a2-be1a-065e152e7df2\") " Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.291571 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sqhg4\" (UniqueName: \"kubernetes.io/projected/77f7b81a-b74e-41a2-be1a-065e152e7df2-kube-api-access-sqhg4\") pod \"77f7b81a-b74e-41a2-be1a-065e152e7df2\" (UID: \"77f7b81a-b74e-41a2-be1a-065e152e7df2\") " Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.291825 4733 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f33ab3b4-cc56-4046-a8c5-bd859c2df2aa-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.291836 4733 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f33ab3b4-cc56-4046-a8c5-bd859c2df2aa-ceph\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.291845 4733 reconciler_common.go:293] "Volume detached for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/f33ab3b4-cc56-4046-a8c5-bd859c2df2aa-var-lib-manila\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.291853 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tzm4m\" (UniqueName: \"kubernetes.io/projected/f33ab3b4-cc56-4046-a8c5-bd859c2df2aa-kube-api-access-tzm4m\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.291862 4733 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f33ab3b4-cc56-4046-a8c5-bd859c2df2aa-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.291870 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f33ab3b4-cc56-4046-a8c5-bd859c2df2aa-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.295090 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77f7b81a-b74e-41a2-be1a-065e152e7df2-utilities" (OuterVolumeSpecName: "utilities") pod "77f7b81a-b74e-41a2-be1a-065e152e7df2" (UID: "77f7b81a-b74e-41a2-be1a-065e152e7df2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.300518 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77f7b81a-b74e-41a2-be1a-065e152e7df2-kube-api-access-sqhg4" (OuterVolumeSpecName: "kube-api-access-sqhg4") pod "77f7b81a-b74e-41a2-be1a-065e152e7df2" (UID: "77f7b81a-b74e-41a2-be1a-065e152e7df2"). InnerVolumeSpecName "kube-api-access-sqhg4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.303993 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f33ab3b4-cc56-4046-a8c5-bd859c2df2aa-config-data" (OuterVolumeSpecName: "config-data") pod "f33ab3b4-cc56-4046-a8c5-bd859c2df2aa" (UID: "f33ab3b4-cc56-4046-a8c5-bd859c2df2aa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.394348 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sqhg4\" (UniqueName: \"kubernetes.io/projected/77f7b81a-b74e-41a2-be1a-065e152e7df2-kube-api-access-sqhg4\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.394377 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77f7b81a-b74e-41a2-be1a-065e152e7df2-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.394389 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f33ab3b4-cc56-4046-a8c5-bd859c2df2aa-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.476252 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77f7b81a-b74e-41a2-be1a-065e152e7df2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "77f7b81a-b74e-41a2-be1a-065e152e7df2" (UID: "77f7b81a-b74e-41a2-be1a-065e152e7df2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.495401 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77f7b81a-b74e-41a2-be1a-065e152e7df2-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.520912 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-64f9574466-56vg6" Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.565199 4733 generic.go:334] "Generic (PLEG): container finished" podID="1ae26a29-79a7-4f0a-b546-f20743ba97a3" containerID="6b98411c753ef28c5622e28730c655cc9b2d9663f5b9c1e89989944c008b5759" exitCode=0 Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.565358 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-index-gqz7t" event={"ID":"1ae26a29-79a7-4f0a-b546-f20743ba97a3","Type":"ContainerDied","Data":"6b98411c753ef28c5622e28730c655cc9b2d9663f5b9c1e89989944c008b5759"} Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.569315 4733 generic.go:334] "Generic (PLEG): container finished" podID="77f7b81a-b74e-41a2-be1a-065e152e7df2" containerID="024ce8576a4037de81134b65c71f1aef864d2bd0300f10e77d3375c981e4dec2" exitCode=0 Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.569393 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zg4qj" event={"ID":"77f7b81a-b74e-41a2-be1a-065e152e7df2","Type":"ContainerDied","Data":"024ce8576a4037de81134b65c71f1aef864d2bd0300f10e77d3375c981e4dec2"} Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.569423 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zg4qj" event={"ID":"77f7b81a-b74e-41a2-be1a-065e152e7df2","Type":"ContainerDied","Data":"1ec2ec1187db4a07b0468376c713004b0eeebf94dbe7c79eb8e4bcb7f84432e1"} Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.569443 4733 scope.go:117] "RemoveContainer" containerID="024ce8576a4037de81134b65c71f1aef864d2bd0300f10e77d3375c981e4dec2" Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.569444 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zg4qj" Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.574389 4733 generic.go:334] "Generic (PLEG): container finished" podID="f3971f39-e65f-4262-acaf-ca5fca3f7cc9" containerID="84e34713078767c3675090908ff02de8038baca92063e381c481b5e26d015c5b" exitCode=0 Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.574547 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-64f9574466-56vg6" Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.574689 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-64f9574466-56vg6" event={"ID":"f3971f39-e65f-4262-acaf-ca5fca3f7cc9","Type":"ContainerDied","Data":"84e34713078767c3675090908ff02de8038baca92063e381c481b5e26d015c5b"} Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.574737 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-64f9574466-56vg6" event={"ID":"f3971f39-e65f-4262-acaf-ca5fca3f7cc9","Type":"ContainerDied","Data":"de176ba5e529933724fa6a4378d07fb016b609f3778e811313906fcbf5442dfd"} Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.581364 4733 generic.go:334] "Generic (PLEG): container finished" podID="9abf907e-85df-4616-a830-56fae1ae3f69" containerID="b7e17883ce2314948330a91af64d17c2235f67db1c6a7c62d5c1097a7e2be57f" exitCode=0 Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.581478 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-scheduler-0" event={"ID":"9abf907e-85df-4616-a830-56fae1ae3f69","Type":"ContainerDied","Data":"b7e17883ce2314948330a91af64d17c2235f67db1c6a7c62d5c1097a7e2be57f"} Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.585994 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manilaa9fc-account-delete-fb8w5" event={"ID":"62b4bf32-8cef-4a0c-bec7-c3511e5f8929","Type":"ContainerStarted","Data":"620e7108713cf105506971c0510c987a106ae4ec3c66f5219886e20d0bc4dbd1"} Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.586043 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manilaa9fc-account-delete-fb8w5" event={"ID":"62b4bf32-8cef-4a0c-bec7-c3511e5f8929","Type":"ContainerStarted","Data":"e8c886cb1cc21b6849729251dbff1cb364495972682272b1a26b5653e87af1b9"} Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.591370 4733 generic.go:334] "Generic (PLEG): container finished" podID="f33ab3b4-cc56-4046-a8c5-bd859c2df2aa" containerID="3423380ed77bc3c40a9c38ef88b85589d65ebcfa3856ed80eb5d3e2993383c89" exitCode=1 Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.591426 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-share-share1-0" event={"ID":"f33ab3b4-cc56-4046-a8c5-bd859c2df2aa","Type":"ContainerDied","Data":"3423380ed77bc3c40a9c38ef88b85589d65ebcfa3856ed80eb5d3e2993383c89"} Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.591452 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-share-share1-0" event={"ID":"f33ab3b4-cc56-4046-a8c5-bd859c2df2aa","Type":"ContainerDied","Data":"db5510313cce5956f185323bca26f48c0b1129157b40cd542257160d19914896"} Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.591479 4733 scope.go:117] "RemoveContainer" containerID="8fe25f86bec5e60c73d0bfaa6c98d148337d99c606b8fcaea3ee0776435e4aab" Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.591588 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-share-share1-0" Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.597500 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f3971f39-e65f-4262-acaf-ca5fca3f7cc9-apiservice-cert\") pod \"f3971f39-e65f-4262-acaf-ca5fca3f7cc9\" (UID: \"f3971f39-e65f-4262-acaf-ca5fca3f7cc9\") " Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.597546 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f3971f39-e65f-4262-acaf-ca5fca3f7cc9-webhook-cert\") pod \"f3971f39-e65f-4262-acaf-ca5fca3f7cc9\" (UID: \"f3971f39-e65f-4262-acaf-ca5fca3f7cc9\") " Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.597610 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mzrxk\" (UniqueName: \"kubernetes.io/projected/f3971f39-e65f-4262-acaf-ca5fca3f7cc9-kube-api-access-mzrxk\") pod \"f3971f39-e65f-4262-acaf-ca5fca3f7cc9\" (UID: \"f3971f39-e65f-4262-acaf-ca5fca3f7cc9\") " Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.598033 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-service-cleanup-n5b5h655-8d2b7" event={"ID":"9dffed5a-a57d-4710-909f-799b47ff68ef","Type":"ContainerStarted","Data":"ad443ac0179f4c8b1942430fb1ee0111ff937d00f7002251606b054c74e1138e"} Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.598177 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="manila-kuttl-tests/manila-service-cleanup-n5b5h655-8d2b7" podUID="9dffed5a-a57d-4710-909f-799b47ff68ef" containerName="manila-service-cleanup-n5b5h655" containerID="cri-o://ad443ac0179f4c8b1942430fb1ee0111ff937d00f7002251606b054c74e1138e" gracePeriod=30 Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.602997 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3971f39-e65f-4262-acaf-ca5fca3f7cc9-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "f3971f39-e65f-4262-acaf-ca5fca3f7cc9" (UID: "f3971f39-e65f-4262-acaf-ca5fca3f7cc9"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.607243 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3971f39-e65f-4262-acaf-ca5fca3f7cc9-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "f3971f39-e65f-4262-acaf-ca5fca3f7cc9" (UID: "f3971f39-e65f-4262-acaf-ca5fca3f7cc9"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.615034 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3971f39-e65f-4262-acaf-ca5fca3f7cc9-kube-api-access-mzrxk" (OuterVolumeSpecName: "kube-api-access-mzrxk") pod "f3971f39-e65f-4262-acaf-ca5fca3f7cc9" (UID: "f3971f39-e65f-4262-acaf-ca5fca3f7cc9"). InnerVolumeSpecName "kube-api-access-mzrxk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.642390 4733 scope.go:117] "RemoveContainer" containerID="216e57641a8246bb773ef10d5a930b48a4e5c7f0a118ea242eee73529c30e36e" Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.645069 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zg4qj"] Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.653805 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-zg4qj"] Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.659292 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="manila-kuttl-tests/manila-service-cleanup-n5b5h655-8d2b7" podStartSLOduration=2.6592715670000002 podStartE2EDuration="2.659271567s" podCreationTimestamp="2026-02-02 15:38:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:38:34.634615896 +0000 UTC m=+1398.086077254" watchObservedRunningTime="2026-02-02 15:38:34.659271567 +0000 UTC m=+1398.110732925" Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.667255 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/manila-share-share1-0"] Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.667836 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-index-gqz7t" Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.674031 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["manila-kuttl-tests/manila-share-share1-0"] Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.677291 4733 scope.go:117] "RemoveContainer" containerID="024ce8576a4037de81134b65c71f1aef864d2bd0300f10e77d3375c981e4dec2" Feb 02 15:38:34 crc kubenswrapper[4733]: E0202 15:38:34.677871 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"024ce8576a4037de81134b65c71f1aef864d2bd0300f10e77d3375c981e4dec2\": container with ID starting with 024ce8576a4037de81134b65c71f1aef864d2bd0300f10e77d3375c981e4dec2 not found: ID does not exist" containerID="024ce8576a4037de81134b65c71f1aef864d2bd0300f10e77d3375c981e4dec2" Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.677897 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"024ce8576a4037de81134b65c71f1aef864d2bd0300f10e77d3375c981e4dec2"} err="failed to get container status \"024ce8576a4037de81134b65c71f1aef864d2bd0300f10e77d3375c981e4dec2\": rpc error: code = NotFound desc = could not find container \"024ce8576a4037de81134b65c71f1aef864d2bd0300f10e77d3375c981e4dec2\": container with ID starting with 024ce8576a4037de81134b65c71f1aef864d2bd0300f10e77d3375c981e4dec2 not found: ID does not exist" Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.677918 4733 scope.go:117] "RemoveContainer" containerID="8fe25f86bec5e60c73d0bfaa6c98d148337d99c606b8fcaea3ee0776435e4aab" Feb 02 15:38:34 crc kubenswrapper[4733]: E0202 15:38:34.678521 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8fe25f86bec5e60c73d0bfaa6c98d148337d99c606b8fcaea3ee0776435e4aab\": container with ID starting with 8fe25f86bec5e60c73d0bfaa6c98d148337d99c606b8fcaea3ee0776435e4aab not found: ID does not exist" containerID="8fe25f86bec5e60c73d0bfaa6c98d148337d99c606b8fcaea3ee0776435e4aab" Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.678574 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8fe25f86bec5e60c73d0bfaa6c98d148337d99c606b8fcaea3ee0776435e4aab"} err="failed to get container status \"8fe25f86bec5e60c73d0bfaa6c98d148337d99c606b8fcaea3ee0776435e4aab\": rpc error: code = NotFound desc = could not find container \"8fe25f86bec5e60c73d0bfaa6c98d148337d99c606b8fcaea3ee0776435e4aab\": container with ID starting with 8fe25f86bec5e60c73d0bfaa6c98d148337d99c606b8fcaea3ee0776435e4aab not found: ID does not exist" Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.678613 4733 scope.go:117] "RemoveContainer" containerID="216e57641a8246bb773ef10d5a930b48a4e5c7f0a118ea242eee73529c30e36e" Feb 02 15:38:34 crc kubenswrapper[4733]: E0202 15:38:34.678998 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"216e57641a8246bb773ef10d5a930b48a4e5c7f0a118ea242eee73529c30e36e\": container with ID starting with 216e57641a8246bb773ef10d5a930b48a4e5c7f0a118ea242eee73529c30e36e not found: ID does not exist" containerID="216e57641a8246bb773ef10d5a930b48a4e5c7f0a118ea242eee73529c30e36e" Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.679022 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"216e57641a8246bb773ef10d5a930b48a4e5c7f0a118ea242eee73529c30e36e"} err="failed to get container status \"216e57641a8246bb773ef10d5a930b48a4e5c7f0a118ea242eee73529c30e36e\": rpc error: code = NotFound desc = could not find container \"216e57641a8246bb773ef10d5a930b48a4e5c7f0a118ea242eee73529c30e36e\": container with ID starting with 216e57641a8246bb773ef10d5a930b48a4e5c7f0a118ea242eee73529c30e36e not found: ID does not exist" Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.679040 4733 scope.go:117] "RemoveContainer" containerID="84e34713078767c3675090908ff02de8038baca92063e381c481b5e26d015c5b" Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.699203 4733 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f3971f39-e65f-4262-acaf-ca5fca3f7cc9-apiservice-cert\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.699250 4733 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f3971f39-e65f-4262-acaf-ca5fca3f7cc9-webhook-cert\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.699264 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mzrxk\" (UniqueName: \"kubernetes.io/projected/f3971f39-e65f-4262-acaf-ca5fca3f7cc9-kube-api-access-mzrxk\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.708062 4733 scope.go:117] "RemoveContainer" containerID="84e34713078767c3675090908ff02de8038baca92063e381c481b5e26d015c5b" Feb 02 15:38:34 crc kubenswrapper[4733]: E0202 15:38:34.708558 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84e34713078767c3675090908ff02de8038baca92063e381c481b5e26d015c5b\": container with ID starting with 84e34713078767c3675090908ff02de8038baca92063e381c481b5e26d015c5b not found: ID does not exist" containerID="84e34713078767c3675090908ff02de8038baca92063e381c481b5e26d015c5b" Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.708596 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84e34713078767c3675090908ff02de8038baca92063e381c481b5e26d015c5b"} err="failed to get container status \"84e34713078767c3675090908ff02de8038baca92063e381c481b5e26d015c5b\": rpc error: code = NotFound desc = could not find container \"84e34713078767c3675090908ff02de8038baca92063e381c481b5e26d015c5b\": container with ID starting with 84e34713078767c3675090908ff02de8038baca92063e381c481b5e26d015c5b not found: ID does not exist" Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.708621 4733 scope.go:117] "RemoveContainer" containerID="f2b8b4d6fef51440a511d27ecdf688ea096171d5dba203231c6cfffcb4880fd6" Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.735370 4733 scope.go:117] "RemoveContainer" containerID="3423380ed77bc3c40a9c38ef88b85589d65ebcfa3856ed80eb5d3e2993383c89" Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.797498 4733 scope.go:117] "RemoveContainer" containerID="f2b8b4d6fef51440a511d27ecdf688ea096171d5dba203231c6cfffcb4880fd6" Feb 02 15:38:34 crc kubenswrapper[4733]: E0202 15:38:34.797988 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2b8b4d6fef51440a511d27ecdf688ea096171d5dba203231c6cfffcb4880fd6\": container with ID starting with f2b8b4d6fef51440a511d27ecdf688ea096171d5dba203231c6cfffcb4880fd6 not found: ID does not exist" containerID="f2b8b4d6fef51440a511d27ecdf688ea096171d5dba203231c6cfffcb4880fd6" Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.798039 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2b8b4d6fef51440a511d27ecdf688ea096171d5dba203231c6cfffcb4880fd6"} err="failed to get container status \"f2b8b4d6fef51440a511d27ecdf688ea096171d5dba203231c6cfffcb4880fd6\": rpc error: code = NotFound desc = could not find container \"f2b8b4d6fef51440a511d27ecdf688ea096171d5dba203231c6cfffcb4880fd6\": container with ID starting with f2b8b4d6fef51440a511d27ecdf688ea096171d5dba203231c6cfffcb4880fd6 not found: ID does not exist" Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.798076 4733 scope.go:117] "RemoveContainer" containerID="3423380ed77bc3c40a9c38ef88b85589d65ebcfa3856ed80eb5d3e2993383c89" Feb 02 15:38:34 crc kubenswrapper[4733]: E0202 15:38:34.798621 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3423380ed77bc3c40a9c38ef88b85589d65ebcfa3856ed80eb5d3e2993383c89\": container with ID starting with 3423380ed77bc3c40a9c38ef88b85589d65ebcfa3856ed80eb5d3e2993383c89 not found: ID does not exist" containerID="3423380ed77bc3c40a9c38ef88b85589d65ebcfa3856ed80eb5d3e2993383c89" Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.798647 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3423380ed77bc3c40a9c38ef88b85589d65ebcfa3856ed80eb5d3e2993383c89"} err="failed to get container status \"3423380ed77bc3c40a9c38ef88b85589d65ebcfa3856ed80eb5d3e2993383c89\": rpc error: code = NotFound desc = could not find container \"3423380ed77bc3c40a9c38ef88b85589d65ebcfa3856ed80eb5d3e2993383c89\": container with ID starting with 3423380ed77bc3c40a9c38ef88b85589d65ebcfa3856ed80eb5d3e2993383c89 not found: ID does not exist" Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.800024 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzgdc\" (UniqueName: \"kubernetes.io/projected/1ae26a29-79a7-4f0a-b546-f20743ba97a3-kube-api-access-fzgdc\") pod \"1ae26a29-79a7-4f0a-b546-f20743ba97a3\" (UID: \"1ae26a29-79a7-4f0a-b546-f20743ba97a3\") " Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.803376 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ae26a29-79a7-4f0a-b546-f20743ba97a3-kube-api-access-fzgdc" (OuterVolumeSpecName: "kube-api-access-fzgdc") pod "1ae26a29-79a7-4f0a-b546-f20743ba97a3" (UID: "1ae26a29-79a7-4f0a-b546-f20743ba97a3"). InnerVolumeSpecName "kube-api-access-fzgdc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.902005 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzgdc\" (UniqueName: \"kubernetes.io/projected/1ae26a29-79a7-4f0a-b546-f20743ba97a3-kube-api-access-fzgdc\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.949794 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/manila-operator-controller-manager-64f9574466-56vg6"] Feb 02 15:38:34 crc kubenswrapper[4733]: I0202 15:38:34.960498 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/manila-operator-controller-manager-64f9574466-56vg6"] Feb 02 15:38:35 crc kubenswrapper[4733]: I0202 15:38:35.147435 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:38:35 crc kubenswrapper[4733]: I0202 15:38:35.269129 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9" path="/var/lib/kubelet/pods/56e9f0c8-ecc5-4af6-87c6-7cbeab57e5f9/volumes" Feb 02 15:38:35 crc kubenswrapper[4733]: I0202 15:38:35.270395 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77f7b81a-b74e-41a2-be1a-065e152e7df2" path="/var/lib/kubelet/pods/77f7b81a-b74e-41a2-be1a-065e152e7df2/volumes" Feb 02 15:38:35 crc kubenswrapper[4733]: I0202 15:38:35.271619 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f33ab3b4-cc56-4046-a8c5-bd859c2df2aa" path="/var/lib/kubelet/pods/f33ab3b4-cc56-4046-a8c5-bd859c2df2aa/volumes" Feb 02 15:38:35 crc kubenswrapper[4733]: I0202 15:38:35.273736 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3971f39-e65f-4262-acaf-ca5fca3f7cc9" path="/var/lib/kubelet/pods/f3971f39-e65f-4262-acaf-ca5fca3f7cc9/volumes" Feb 02 15:38:35 crc kubenswrapper[4733]: I0202 15:38:35.306997 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8gmm6\" (UniqueName: \"kubernetes.io/projected/9abf907e-85df-4616-a830-56fae1ae3f69-kube-api-access-8gmm6\") pod \"9abf907e-85df-4616-a830-56fae1ae3f69\" (UID: \"9abf907e-85df-4616-a830-56fae1ae3f69\") " Feb 02 15:38:35 crc kubenswrapper[4733]: I0202 15:38:35.307103 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9abf907e-85df-4616-a830-56fae1ae3f69-etc-machine-id\") pod \"9abf907e-85df-4616-a830-56fae1ae3f69\" (UID: \"9abf907e-85df-4616-a830-56fae1ae3f69\") " Feb 02 15:38:35 crc kubenswrapper[4733]: I0202 15:38:35.307257 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9abf907e-85df-4616-a830-56fae1ae3f69-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "9abf907e-85df-4616-a830-56fae1ae3f69" (UID: "9abf907e-85df-4616-a830-56fae1ae3f69"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 15:38:35 crc kubenswrapper[4733]: I0202 15:38:35.307307 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9abf907e-85df-4616-a830-56fae1ae3f69-config-data-custom\") pod \"9abf907e-85df-4616-a830-56fae1ae3f69\" (UID: \"9abf907e-85df-4616-a830-56fae1ae3f69\") " Feb 02 15:38:35 crc kubenswrapper[4733]: I0202 15:38:35.307374 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9abf907e-85df-4616-a830-56fae1ae3f69-config-data\") pod \"9abf907e-85df-4616-a830-56fae1ae3f69\" (UID: \"9abf907e-85df-4616-a830-56fae1ae3f69\") " Feb 02 15:38:35 crc kubenswrapper[4733]: I0202 15:38:35.307464 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9abf907e-85df-4616-a830-56fae1ae3f69-scripts\") pod \"9abf907e-85df-4616-a830-56fae1ae3f69\" (UID: \"9abf907e-85df-4616-a830-56fae1ae3f69\") " Feb 02 15:38:35 crc kubenswrapper[4733]: I0202 15:38:35.307859 4733 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9abf907e-85df-4616-a830-56fae1ae3f69-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:35 crc kubenswrapper[4733]: I0202 15:38:35.311759 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9abf907e-85df-4616-a830-56fae1ae3f69-scripts" (OuterVolumeSpecName: "scripts") pod "9abf907e-85df-4616-a830-56fae1ae3f69" (UID: "9abf907e-85df-4616-a830-56fae1ae3f69"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:38:35 crc kubenswrapper[4733]: I0202 15:38:35.312351 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9abf907e-85df-4616-a830-56fae1ae3f69-kube-api-access-8gmm6" (OuterVolumeSpecName: "kube-api-access-8gmm6") pod "9abf907e-85df-4616-a830-56fae1ae3f69" (UID: "9abf907e-85df-4616-a830-56fae1ae3f69"). InnerVolumeSpecName "kube-api-access-8gmm6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:38:35 crc kubenswrapper[4733]: I0202 15:38:35.312954 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9abf907e-85df-4616-a830-56fae1ae3f69-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "9abf907e-85df-4616-a830-56fae1ae3f69" (UID: "9abf907e-85df-4616-a830-56fae1ae3f69"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:38:35 crc kubenswrapper[4733]: I0202 15:38:35.369074 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9abf907e-85df-4616-a830-56fae1ae3f69-config-data" (OuterVolumeSpecName: "config-data") pod "9abf907e-85df-4616-a830-56fae1ae3f69" (UID: "9abf907e-85df-4616-a830-56fae1ae3f69"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:38:35 crc kubenswrapper[4733]: I0202 15:38:35.408808 4733 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9abf907e-85df-4616-a830-56fae1ae3f69-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:35 crc kubenswrapper[4733]: I0202 15:38:35.408844 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9abf907e-85df-4616-a830-56fae1ae3f69-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:35 crc kubenswrapper[4733]: I0202 15:38:35.408853 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9abf907e-85df-4616-a830-56fae1ae3f69-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:35 crc kubenswrapper[4733]: I0202 15:38:35.408862 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8gmm6\" (UniqueName: \"kubernetes.io/projected/9abf907e-85df-4616-a830-56fae1ae3f69-kube-api-access-8gmm6\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:35 crc kubenswrapper[4733]: I0202 15:38:35.608623 4733 generic.go:334] "Generic (PLEG): container finished" podID="62b4bf32-8cef-4a0c-bec7-c3511e5f8929" containerID="620e7108713cf105506971c0510c987a106ae4ec3c66f5219886e20d0bc4dbd1" exitCode=0 Feb 02 15:38:35 crc kubenswrapper[4733]: I0202 15:38:35.608687 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manilaa9fc-account-delete-fb8w5" event={"ID":"62b4bf32-8cef-4a0c-bec7-c3511e5f8929","Type":"ContainerDied","Data":"620e7108713cf105506971c0510c987a106ae4ec3c66f5219886e20d0bc4dbd1"} Feb 02 15:38:35 crc kubenswrapper[4733]: I0202 15:38:35.613111 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-index-gqz7t" Feb 02 15:38:35 crc kubenswrapper[4733]: I0202 15:38:35.614009 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-index-gqz7t" event={"ID":"1ae26a29-79a7-4f0a-b546-f20743ba97a3","Type":"ContainerDied","Data":"51f3f474f991a56c8b947253da5d206e93197e2fee3103d564ddef69beadf504"} Feb 02 15:38:35 crc kubenswrapper[4733]: I0202 15:38:35.614055 4733 scope.go:117] "RemoveContainer" containerID="6b98411c753ef28c5622e28730c655cc9b2d9663f5b9c1e89989944c008b5759" Feb 02 15:38:35 crc kubenswrapper[4733]: I0202 15:38:35.630305 4733 generic.go:334] "Generic (PLEG): container finished" podID="9abf907e-85df-4616-a830-56fae1ae3f69" containerID="1b4db59339def5c0775f38a97cbcb2c76ad7e0dde03ca35dd3b89b993012719d" exitCode=0 Feb 02 15:38:35 crc kubenswrapper[4733]: I0202 15:38:35.630358 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-scheduler-0" event={"ID":"9abf907e-85df-4616-a830-56fae1ae3f69","Type":"ContainerDied","Data":"1b4db59339def5c0775f38a97cbcb2c76ad7e0dde03ca35dd3b89b993012719d"} Feb 02 15:38:35 crc kubenswrapper[4733]: I0202 15:38:35.630406 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-scheduler-0" event={"ID":"9abf907e-85df-4616-a830-56fae1ae3f69","Type":"ContainerDied","Data":"55cf3d26df3efaf1b447714e8d36cb6bdcb9c9307b451802531a312133d09ad0"} Feb 02 15:38:35 crc kubenswrapper[4733]: I0202 15:38:35.630438 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-scheduler-0" Feb 02 15:38:35 crc kubenswrapper[4733]: I0202 15:38:35.643968 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/manila-operator-index-gqz7t"] Feb 02 15:38:35 crc kubenswrapper[4733]: I0202 15:38:35.655216 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/manila-operator-index-gqz7t"] Feb 02 15:38:35 crc kubenswrapper[4733]: I0202 15:38:35.659715 4733 scope.go:117] "RemoveContainer" containerID="b7e17883ce2314948330a91af64d17c2235f67db1c6a7c62d5c1097a7e2be57f" Feb 02 15:38:35 crc kubenswrapper[4733]: I0202 15:38:35.677633 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/manila-scheduler-0"] Feb 02 15:38:35 crc kubenswrapper[4733]: I0202 15:38:35.681536 4733 scope.go:117] "RemoveContainer" containerID="1b4db59339def5c0775f38a97cbcb2c76ad7e0dde03ca35dd3b89b993012719d" Feb 02 15:38:35 crc kubenswrapper[4733]: I0202 15:38:35.683461 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["manila-kuttl-tests/manila-scheduler-0"] Feb 02 15:38:35 crc kubenswrapper[4733]: I0202 15:38:35.699133 4733 scope.go:117] "RemoveContainer" containerID="b7e17883ce2314948330a91af64d17c2235f67db1c6a7c62d5c1097a7e2be57f" Feb 02 15:38:35 crc kubenswrapper[4733]: E0202 15:38:35.699581 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7e17883ce2314948330a91af64d17c2235f67db1c6a7c62d5c1097a7e2be57f\": container with ID starting with b7e17883ce2314948330a91af64d17c2235f67db1c6a7c62d5c1097a7e2be57f not found: ID does not exist" containerID="b7e17883ce2314948330a91af64d17c2235f67db1c6a7c62d5c1097a7e2be57f" Feb 02 15:38:35 crc kubenswrapper[4733]: I0202 15:38:35.699612 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7e17883ce2314948330a91af64d17c2235f67db1c6a7c62d5c1097a7e2be57f"} err="failed to get container status \"b7e17883ce2314948330a91af64d17c2235f67db1c6a7c62d5c1097a7e2be57f\": rpc error: code = NotFound desc = could not find container \"b7e17883ce2314948330a91af64d17c2235f67db1c6a7c62d5c1097a7e2be57f\": container with ID starting with b7e17883ce2314948330a91af64d17c2235f67db1c6a7c62d5c1097a7e2be57f not found: ID does not exist" Feb 02 15:38:35 crc kubenswrapper[4733]: I0202 15:38:35.699632 4733 scope.go:117] "RemoveContainer" containerID="1b4db59339def5c0775f38a97cbcb2c76ad7e0dde03ca35dd3b89b993012719d" Feb 02 15:38:35 crc kubenswrapper[4733]: E0202 15:38:35.699887 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b4db59339def5c0775f38a97cbcb2c76ad7e0dde03ca35dd3b89b993012719d\": container with ID starting with 1b4db59339def5c0775f38a97cbcb2c76ad7e0dde03ca35dd3b89b993012719d not found: ID does not exist" containerID="1b4db59339def5c0775f38a97cbcb2c76ad7e0dde03ca35dd3b89b993012719d" Feb 02 15:38:35 crc kubenswrapper[4733]: I0202 15:38:35.699918 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b4db59339def5c0775f38a97cbcb2c76ad7e0dde03ca35dd3b89b993012719d"} err="failed to get container status \"1b4db59339def5c0775f38a97cbcb2c76ad7e0dde03ca35dd3b89b993012719d\": rpc error: code = NotFound desc = could not find container \"1b4db59339def5c0775f38a97cbcb2c76ad7e0dde03ca35dd3b89b993012719d\": container with ID starting with 1b4db59339def5c0775f38a97cbcb2c76ad7e0dde03ca35dd3b89b993012719d not found: ID does not exist" Feb 02 15:38:35 crc kubenswrapper[4733]: I0202 15:38:35.896762 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manilaa9fc-account-delete-fb8w5" Feb 02 15:38:36 crc kubenswrapper[4733]: I0202 15:38:36.016503 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ppl9m\" (UniqueName: \"kubernetes.io/projected/62b4bf32-8cef-4a0c-bec7-c3511e5f8929-kube-api-access-ppl9m\") pod \"62b4bf32-8cef-4a0c-bec7-c3511e5f8929\" (UID: \"62b4bf32-8cef-4a0c-bec7-c3511e5f8929\") " Feb 02 15:38:36 crc kubenswrapper[4733]: I0202 15:38:36.016728 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/62b4bf32-8cef-4a0c-bec7-c3511e5f8929-operator-scripts\") pod \"62b4bf32-8cef-4a0c-bec7-c3511e5f8929\" (UID: \"62b4bf32-8cef-4a0c-bec7-c3511e5f8929\") " Feb 02 15:38:36 crc kubenswrapper[4733]: I0202 15:38:36.017356 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62b4bf32-8cef-4a0c-bec7-c3511e5f8929-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "62b4bf32-8cef-4a0c-bec7-c3511e5f8929" (UID: "62b4bf32-8cef-4a0c-bec7-c3511e5f8929"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:38:36 crc kubenswrapper[4733]: I0202 15:38:36.024814 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62b4bf32-8cef-4a0c-bec7-c3511e5f8929-kube-api-access-ppl9m" (OuterVolumeSpecName: "kube-api-access-ppl9m") pod "62b4bf32-8cef-4a0c-bec7-c3511e5f8929" (UID: "62b4bf32-8cef-4a0c-bec7-c3511e5f8929"). InnerVolumeSpecName "kube-api-access-ppl9m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:38:36 crc kubenswrapper[4733]: I0202 15:38:36.118766 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ppl9m\" (UniqueName: \"kubernetes.io/projected/62b4bf32-8cef-4a0c-bec7-c3511e5f8929-kube-api-access-ppl9m\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:36 crc kubenswrapper[4733]: I0202 15:38:36.118828 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/62b4bf32-8cef-4a0c-bec7-c3511e5f8929-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:36 crc kubenswrapper[4733]: I0202 15:38:36.648586 4733 generic.go:334] "Generic (PLEG): container finished" podID="fe12e7c9-118b-4413-8fa4-96745802310e" containerID="0c0517983e8f69418d96a083e721d09d7185d0fbb013e0b4fc4e93c6e1b2a795" exitCode=0 Feb 02 15:38:36 crc kubenswrapper[4733]: I0202 15:38:36.648669 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-api-0" event={"ID":"fe12e7c9-118b-4413-8fa4-96745802310e","Type":"ContainerDied","Data":"0c0517983e8f69418d96a083e721d09d7185d0fbb013e0b4fc4e93c6e1b2a795"} Feb 02 15:38:36 crc kubenswrapper[4733]: I0202 15:38:36.652699 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manilaa9fc-account-delete-fb8w5" event={"ID":"62b4bf32-8cef-4a0c-bec7-c3511e5f8929","Type":"ContainerDied","Data":"e8c886cb1cc21b6849729251dbff1cb364495972682272b1a26b5653e87af1b9"} Feb 02 15:38:36 crc kubenswrapper[4733]: I0202 15:38:36.652740 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e8c886cb1cc21b6849729251dbff1cb364495972682272b1a26b5653e87af1b9" Feb 02 15:38:36 crc kubenswrapper[4733]: I0202 15:38:36.652757 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manilaa9fc-account-delete-fb8w5" Feb 02 15:38:36 crc kubenswrapper[4733]: I0202 15:38:36.660508 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-api-0" Feb 02 15:38:36 crc kubenswrapper[4733]: I0202 15:38:36.726306 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dmmmb\" (UniqueName: \"kubernetes.io/projected/fe12e7c9-118b-4413-8fa4-96745802310e-kube-api-access-dmmmb\") pod \"fe12e7c9-118b-4413-8fa4-96745802310e\" (UID: \"fe12e7c9-118b-4413-8fa4-96745802310e\") " Feb 02 15:38:36 crc kubenswrapper[4733]: I0202 15:38:36.726388 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe12e7c9-118b-4413-8fa4-96745802310e-config-data\") pod \"fe12e7c9-118b-4413-8fa4-96745802310e\" (UID: \"fe12e7c9-118b-4413-8fa4-96745802310e\") " Feb 02 15:38:36 crc kubenswrapper[4733]: I0202 15:38:36.726427 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe12e7c9-118b-4413-8fa4-96745802310e-scripts\") pod \"fe12e7c9-118b-4413-8fa4-96745802310e\" (UID: \"fe12e7c9-118b-4413-8fa4-96745802310e\") " Feb 02 15:38:36 crc kubenswrapper[4733]: I0202 15:38:36.726462 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fe12e7c9-118b-4413-8fa4-96745802310e-config-data-custom\") pod \"fe12e7c9-118b-4413-8fa4-96745802310e\" (UID: \"fe12e7c9-118b-4413-8fa4-96745802310e\") " Feb 02 15:38:36 crc kubenswrapper[4733]: I0202 15:38:36.726546 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fe12e7c9-118b-4413-8fa4-96745802310e-logs\") pod \"fe12e7c9-118b-4413-8fa4-96745802310e\" (UID: \"fe12e7c9-118b-4413-8fa4-96745802310e\") " Feb 02 15:38:36 crc kubenswrapper[4733]: I0202 15:38:36.726596 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fe12e7c9-118b-4413-8fa4-96745802310e-etc-machine-id\") pod \"fe12e7c9-118b-4413-8fa4-96745802310e\" (UID: \"fe12e7c9-118b-4413-8fa4-96745802310e\") " Feb 02 15:38:36 crc kubenswrapper[4733]: I0202 15:38:36.726889 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fe12e7c9-118b-4413-8fa4-96745802310e-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "fe12e7c9-118b-4413-8fa4-96745802310e" (UID: "fe12e7c9-118b-4413-8fa4-96745802310e"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 02 15:38:36 crc kubenswrapper[4733]: I0202 15:38:36.727046 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe12e7c9-118b-4413-8fa4-96745802310e-logs" (OuterVolumeSpecName: "logs") pod "fe12e7c9-118b-4413-8fa4-96745802310e" (UID: "fe12e7c9-118b-4413-8fa4-96745802310e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:38:36 crc kubenswrapper[4733]: I0202 15:38:36.730155 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe12e7c9-118b-4413-8fa4-96745802310e-scripts" (OuterVolumeSpecName: "scripts") pod "fe12e7c9-118b-4413-8fa4-96745802310e" (UID: "fe12e7c9-118b-4413-8fa4-96745802310e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:38:36 crc kubenswrapper[4733]: I0202 15:38:36.730201 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe12e7c9-118b-4413-8fa4-96745802310e-kube-api-access-dmmmb" (OuterVolumeSpecName: "kube-api-access-dmmmb") pod "fe12e7c9-118b-4413-8fa4-96745802310e" (UID: "fe12e7c9-118b-4413-8fa4-96745802310e"). InnerVolumeSpecName "kube-api-access-dmmmb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:38:36 crc kubenswrapper[4733]: I0202 15:38:36.730620 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe12e7c9-118b-4413-8fa4-96745802310e-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "fe12e7c9-118b-4413-8fa4-96745802310e" (UID: "fe12e7c9-118b-4413-8fa4-96745802310e"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:38:36 crc kubenswrapper[4733]: I0202 15:38:36.762399 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe12e7c9-118b-4413-8fa4-96745802310e-config-data" (OuterVolumeSpecName: "config-data") pod "fe12e7c9-118b-4413-8fa4-96745802310e" (UID: "fe12e7c9-118b-4413-8fa4-96745802310e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:38:36 crc kubenswrapper[4733]: I0202 15:38:36.828069 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe12e7c9-118b-4413-8fa4-96745802310e-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:36 crc kubenswrapper[4733]: I0202 15:38:36.828275 4733 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fe12e7c9-118b-4413-8fa4-96745802310e-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:36 crc kubenswrapper[4733]: I0202 15:38:36.828333 4733 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fe12e7c9-118b-4413-8fa4-96745802310e-logs\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:36 crc kubenswrapper[4733]: I0202 15:38:36.828419 4733 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fe12e7c9-118b-4413-8fa4-96745802310e-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:36 crc kubenswrapper[4733]: I0202 15:38:36.828471 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dmmmb\" (UniqueName: \"kubernetes.io/projected/fe12e7c9-118b-4413-8fa4-96745802310e-kube-api-access-dmmmb\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:36 crc kubenswrapper[4733]: I0202 15:38:36.828520 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe12e7c9-118b-4413-8fa4-96745802310e-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:37 crc kubenswrapper[4733]: I0202 15:38:37.267421 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ae26a29-79a7-4f0a-b546-f20743ba97a3" path="/var/lib/kubelet/pods/1ae26a29-79a7-4f0a-b546-f20743ba97a3/volumes" Feb 02 15:38:37 crc kubenswrapper[4733]: I0202 15:38:37.268424 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9abf907e-85df-4616-a830-56fae1ae3f69" path="/var/lib/kubelet/pods/9abf907e-85df-4616-a830-56fae1ae3f69/volumes" Feb 02 15:38:37 crc kubenswrapper[4733]: I0202 15:38:37.665223 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-api-0" event={"ID":"fe12e7c9-118b-4413-8fa4-96745802310e","Type":"ContainerDied","Data":"4bae261d9bc30895052af22344146c6928e78f18b17985b7a57f92c4f560e7d3"} Feb 02 15:38:37 crc kubenswrapper[4733]: I0202 15:38:37.665288 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-api-0" Feb 02 15:38:37 crc kubenswrapper[4733]: I0202 15:38:37.665313 4733 scope.go:117] "RemoveContainer" containerID="0c0517983e8f69418d96a083e721d09d7185d0fbb013e0b4fc4e93c6e1b2a795" Feb 02 15:38:37 crc kubenswrapper[4733]: I0202 15:38:37.701146 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/manila-api-0"] Feb 02 15:38:37 crc kubenswrapper[4733]: I0202 15:38:37.703060 4733 scope.go:117] "RemoveContainer" containerID="11acc8fca457138b1b74f4ee6a9396a9e163b6a49c60cad0a5e39487d7377acf" Feb 02 15:38:37 crc kubenswrapper[4733]: I0202 15:38:37.711006 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["manila-kuttl-tests/manila-api-0"] Feb 02 15:38:38 crc kubenswrapper[4733]: I0202 15:38:38.189894 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/manila-db-create-jwlhx"] Feb 02 15:38:38 crc kubenswrapper[4733]: I0202 15:38:38.199699 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["manila-kuttl-tests/manila-db-create-jwlhx"] Feb 02 15:38:38 crc kubenswrapper[4733]: I0202 15:38:38.206848 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/manilaa9fc-account-delete-fb8w5"] Feb 02 15:38:38 crc kubenswrapper[4733]: I0202 15:38:38.213331 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["manila-kuttl-tests/manilaa9fc-account-delete-fb8w5"] Feb 02 15:38:38 crc kubenswrapper[4733]: I0202 15:38:38.218764 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/manila-a9fc-account-create-update-9mw4z"] Feb 02 15:38:38 crc kubenswrapper[4733]: I0202 15:38:38.222981 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["manila-kuttl-tests/manila-a9fc-account-create-update-9mw4z"] Feb 02 15:38:39 crc kubenswrapper[4733]: I0202 15:38:39.267633 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b0e265c-62db-493c-9385-863c39c53ed6" path="/var/lib/kubelet/pods/5b0e265c-62db-493c-9385-863c39c53ed6/volumes" Feb 02 15:38:39 crc kubenswrapper[4733]: I0202 15:38:39.269097 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62b4bf32-8cef-4a0c-bec7-c3511e5f8929" path="/var/lib/kubelet/pods/62b4bf32-8cef-4a0c-bec7-c3511e5f8929/volumes" Feb 02 15:38:39 crc kubenswrapper[4733]: I0202 15:38:39.270144 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef267821-af74-4e8a-bcd6-216f35b43880" path="/var/lib/kubelet/pods/ef267821-af74-4e8a-bcd6-216f35b43880/volumes" Feb 02 15:38:39 crc kubenswrapper[4733]: I0202 15:38:39.272245 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe12e7c9-118b-4413-8fa4-96745802310e" path="/var/lib/kubelet/pods/fe12e7c9-118b-4413-8fa4-96745802310e/volumes" Feb 02 15:38:40 crc kubenswrapper[4733]: I0202 15:38:40.033925 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/keystone-db-sync-7wbzb"] Feb 02 15:38:40 crc kubenswrapper[4733]: I0202 15:38:40.043220 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["manila-kuttl-tests/keystone-db-sync-7wbzb"] Feb 02 15:38:40 crc kubenswrapper[4733]: I0202 15:38:40.043947 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/keystone-bootstrap-65qw9"] Feb 02 15:38:40 crc kubenswrapper[4733]: I0202 15:38:40.051671 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["manila-kuttl-tests/keystone-bootstrap-65qw9"] Feb 02 15:38:40 crc kubenswrapper[4733]: I0202 15:38:40.069942 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/keystone-5f5c86bdc8-jpphg"] Feb 02 15:38:40 crc kubenswrapper[4733]: I0202 15:38:40.070318 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="manila-kuttl-tests/keystone-5f5c86bdc8-jpphg" podUID="a972e8cd-cae8-4894-b2d3-7bfff75acc1c" containerName="keystone-api" containerID="cri-o://818ce15b55c28e765033c326ced7cc8bb1dd7b47fe329b86dc8b9830026fccea" gracePeriod=30 Feb 02 15:38:40 crc kubenswrapper[4733]: I0202 15:38:40.090463 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["manila-kuttl-tests/keystonec686-account-delete-zqss2"] Feb 02 15:38:40 crc kubenswrapper[4733]: E0202 15:38:40.090777 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ae26a29-79a7-4f0a-b546-f20743ba97a3" containerName="registry-server" Feb 02 15:38:40 crc kubenswrapper[4733]: I0202 15:38:40.090793 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ae26a29-79a7-4f0a-b546-f20743ba97a3" containerName="registry-server" Feb 02 15:38:40 crc kubenswrapper[4733]: E0202 15:38:40.090806 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f33ab3b4-cc56-4046-a8c5-bd859c2df2aa" containerName="probe" Feb 02 15:38:40 crc kubenswrapper[4733]: I0202 15:38:40.090815 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f33ab3b4-cc56-4046-a8c5-bd859c2df2aa" containerName="probe" Feb 02 15:38:40 crc kubenswrapper[4733]: E0202 15:38:40.090826 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f33ab3b4-cc56-4046-a8c5-bd859c2df2aa" containerName="manila-share" Feb 02 15:38:40 crc kubenswrapper[4733]: I0202 15:38:40.090833 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f33ab3b4-cc56-4046-a8c5-bd859c2df2aa" containerName="manila-share" Feb 02 15:38:40 crc kubenswrapper[4733]: E0202 15:38:40.090846 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe12e7c9-118b-4413-8fa4-96745802310e" containerName="manila-api" Feb 02 15:38:40 crc kubenswrapper[4733]: I0202 15:38:40.090853 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe12e7c9-118b-4413-8fa4-96745802310e" containerName="manila-api" Feb 02 15:38:40 crc kubenswrapper[4733]: E0202 15:38:40.090870 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe12e7c9-118b-4413-8fa4-96745802310e" containerName="manila-api-log" Feb 02 15:38:40 crc kubenswrapper[4733]: I0202 15:38:40.090878 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe12e7c9-118b-4413-8fa4-96745802310e" containerName="manila-api-log" Feb 02 15:38:40 crc kubenswrapper[4733]: E0202 15:38:40.090889 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9abf907e-85df-4616-a830-56fae1ae3f69" containerName="probe" Feb 02 15:38:40 crc kubenswrapper[4733]: I0202 15:38:40.090897 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="9abf907e-85df-4616-a830-56fae1ae3f69" containerName="probe" Feb 02 15:38:40 crc kubenswrapper[4733]: E0202 15:38:40.090916 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77f7b81a-b74e-41a2-be1a-065e152e7df2" containerName="extract-utilities" Feb 02 15:38:40 crc kubenswrapper[4733]: I0202 15:38:40.090925 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="77f7b81a-b74e-41a2-be1a-065e152e7df2" containerName="extract-utilities" Feb 02 15:38:40 crc kubenswrapper[4733]: E0202 15:38:40.090936 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9abf907e-85df-4616-a830-56fae1ae3f69" containerName="manila-scheduler" Feb 02 15:38:40 crc kubenswrapper[4733]: I0202 15:38:40.090943 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="9abf907e-85df-4616-a830-56fae1ae3f69" containerName="manila-scheduler" Feb 02 15:38:40 crc kubenswrapper[4733]: E0202 15:38:40.090951 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77f7b81a-b74e-41a2-be1a-065e152e7df2" containerName="registry-server" Feb 02 15:38:40 crc kubenswrapper[4733]: I0202 15:38:40.090958 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="77f7b81a-b74e-41a2-be1a-065e152e7df2" containerName="registry-server" Feb 02 15:38:40 crc kubenswrapper[4733]: E0202 15:38:40.090973 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3971f39-e65f-4262-acaf-ca5fca3f7cc9" containerName="manager" Feb 02 15:38:40 crc kubenswrapper[4733]: I0202 15:38:40.090981 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3971f39-e65f-4262-acaf-ca5fca3f7cc9" containerName="manager" Feb 02 15:38:40 crc kubenswrapper[4733]: E0202 15:38:40.090992 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77f7b81a-b74e-41a2-be1a-065e152e7df2" containerName="extract-content" Feb 02 15:38:40 crc kubenswrapper[4733]: I0202 15:38:40.091000 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="77f7b81a-b74e-41a2-be1a-065e152e7df2" containerName="extract-content" Feb 02 15:38:40 crc kubenswrapper[4733]: E0202 15:38:40.091013 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62b4bf32-8cef-4a0c-bec7-c3511e5f8929" containerName="mariadb-account-delete" Feb 02 15:38:40 crc kubenswrapper[4733]: I0202 15:38:40.091021 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="62b4bf32-8cef-4a0c-bec7-c3511e5f8929" containerName="mariadb-account-delete" Feb 02 15:38:40 crc kubenswrapper[4733]: I0202 15:38:40.091152 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f33ab3b4-cc56-4046-a8c5-bd859c2df2aa" containerName="probe" Feb 02 15:38:40 crc kubenswrapper[4733]: I0202 15:38:40.091183 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe12e7c9-118b-4413-8fa4-96745802310e" containerName="manila-api" Feb 02 15:38:40 crc kubenswrapper[4733]: I0202 15:38:40.091193 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3971f39-e65f-4262-acaf-ca5fca3f7cc9" containerName="manager" Feb 02 15:38:40 crc kubenswrapper[4733]: I0202 15:38:40.091208 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="77f7b81a-b74e-41a2-be1a-065e152e7df2" containerName="registry-server" Feb 02 15:38:40 crc kubenswrapper[4733]: I0202 15:38:40.091218 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="9abf907e-85df-4616-a830-56fae1ae3f69" containerName="manila-scheduler" Feb 02 15:38:40 crc kubenswrapper[4733]: I0202 15:38:40.091230 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f33ab3b4-cc56-4046-a8c5-bd859c2df2aa" containerName="manila-share" Feb 02 15:38:40 crc kubenswrapper[4733]: I0202 15:38:40.091239 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe12e7c9-118b-4413-8fa4-96745802310e" containerName="manila-api-log" Feb 02 15:38:40 crc kubenswrapper[4733]: I0202 15:38:40.091247 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="62b4bf32-8cef-4a0c-bec7-c3511e5f8929" containerName="mariadb-account-delete" Feb 02 15:38:40 crc kubenswrapper[4733]: I0202 15:38:40.091261 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="9abf907e-85df-4616-a830-56fae1ae3f69" containerName="probe" Feb 02 15:38:40 crc kubenswrapper[4733]: I0202 15:38:40.091271 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ae26a29-79a7-4f0a-b546-f20743ba97a3" containerName="registry-server" Feb 02 15:38:40 crc kubenswrapper[4733]: I0202 15:38:40.091782 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/keystonec686-account-delete-zqss2" Feb 02 15:38:40 crc kubenswrapper[4733]: I0202 15:38:40.112244 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/keystonec686-account-delete-zqss2"] Feb 02 15:38:40 crc kubenswrapper[4733]: I0202 15:38:40.186584 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/79f9eeab-dccd-4404-8179-fc3f776d8b22-operator-scripts\") pod \"keystonec686-account-delete-zqss2\" (UID: \"79f9eeab-dccd-4404-8179-fc3f776d8b22\") " pod="manila-kuttl-tests/keystonec686-account-delete-zqss2" Feb 02 15:38:40 crc kubenswrapper[4733]: I0202 15:38:40.186737 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mrmp\" (UniqueName: \"kubernetes.io/projected/79f9eeab-dccd-4404-8179-fc3f776d8b22-kube-api-access-2mrmp\") pod \"keystonec686-account-delete-zqss2\" (UID: \"79f9eeab-dccd-4404-8179-fc3f776d8b22\") " pod="manila-kuttl-tests/keystonec686-account-delete-zqss2" Feb 02 15:38:40 crc kubenswrapper[4733]: I0202 15:38:40.287847 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mrmp\" (UniqueName: \"kubernetes.io/projected/79f9eeab-dccd-4404-8179-fc3f776d8b22-kube-api-access-2mrmp\") pod \"keystonec686-account-delete-zqss2\" (UID: \"79f9eeab-dccd-4404-8179-fc3f776d8b22\") " pod="manila-kuttl-tests/keystonec686-account-delete-zqss2" Feb 02 15:38:40 crc kubenswrapper[4733]: I0202 15:38:40.287967 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/79f9eeab-dccd-4404-8179-fc3f776d8b22-operator-scripts\") pod \"keystonec686-account-delete-zqss2\" (UID: \"79f9eeab-dccd-4404-8179-fc3f776d8b22\") " pod="manila-kuttl-tests/keystonec686-account-delete-zqss2" Feb 02 15:38:40 crc kubenswrapper[4733]: I0202 15:38:40.289126 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/79f9eeab-dccd-4404-8179-fc3f776d8b22-operator-scripts\") pod \"keystonec686-account-delete-zqss2\" (UID: \"79f9eeab-dccd-4404-8179-fc3f776d8b22\") " pod="manila-kuttl-tests/keystonec686-account-delete-zqss2" Feb 02 15:38:40 crc kubenswrapper[4733]: I0202 15:38:40.309716 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mrmp\" (UniqueName: \"kubernetes.io/projected/79f9eeab-dccd-4404-8179-fc3f776d8b22-kube-api-access-2mrmp\") pod \"keystonec686-account-delete-zqss2\" (UID: \"79f9eeab-dccd-4404-8179-fc3f776d8b22\") " pod="manila-kuttl-tests/keystonec686-account-delete-zqss2" Feb 02 15:38:40 crc kubenswrapper[4733]: I0202 15:38:40.424000 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/keystonec686-account-delete-zqss2" Feb 02 15:38:40 crc kubenswrapper[4733]: I0202 15:38:40.749432 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/root-account-create-update-4jtql"] Feb 02 15:38:40 crc kubenswrapper[4733]: I0202 15:38:40.759504 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["manila-kuttl-tests/root-account-create-update-4jtql"] Feb 02 15:38:40 crc kubenswrapper[4733]: I0202 15:38:40.783390 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/openstack-galera-1"] Feb 02 15:38:40 crc kubenswrapper[4733]: I0202 15:38:40.788236 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/openstack-galera-0"] Feb 02 15:38:40 crc kubenswrapper[4733]: I0202 15:38:40.793136 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/openstack-galera-2"] Feb 02 15:38:40 crc kubenswrapper[4733]: I0202 15:38:40.894520 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/keystonec686-account-delete-zqss2"] Feb 02 15:38:40 crc kubenswrapper[4733]: I0202 15:38:40.958211 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="manila-kuttl-tests/openstack-galera-2" podUID="02261047-2e53-4b34-bde5-5da7facb243c" containerName="galera" containerID="cri-o://8a9cc594f867e17e58e08d3984767170690347ae2f35b2e53796aa7c6d0db5f2" gracePeriod=30 Feb 02 15:38:41 crc kubenswrapper[4733]: I0202 15:38:41.282739 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="158ef91e-84f7-4e1e-a476-569c8a755226" path="/var/lib/kubelet/pods/158ef91e-84f7-4e1e-a476-569c8a755226/volumes" Feb 02 15:38:41 crc kubenswrapper[4733]: I0202 15:38:41.284063 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21e8b34c-9313-4546-8c0d-0b3b014e190c" path="/var/lib/kubelet/pods/21e8b34c-9313-4546-8c0d-0b3b014e190c/volumes" Feb 02 15:38:41 crc kubenswrapper[4733]: I0202 15:38:41.285210 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f36bcd25-3798-4fbf-b317-b6d52b53af05" path="/var/lib/kubelet/pods/f36bcd25-3798-4fbf-b317-b6d52b53af05/volumes" Feb 02 15:38:41 crc kubenswrapper[4733]: I0202 15:38:41.365899 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/memcached-0"] Feb 02 15:38:41 crc kubenswrapper[4733]: I0202 15:38:41.366142 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="manila-kuttl-tests/memcached-0" podUID="e8f2d3a6-ad9f-4e1c-8f9a-41f8c7efcdac" containerName="memcached" containerID="cri-o://60f342505fd409f05a577cf0a9883bd43822c387690a1b4617f376e8acec2dc4" gracePeriod=30 Feb 02 15:38:41 crc kubenswrapper[4733]: I0202 15:38:41.703669 4733 generic.go:334] "Generic (PLEG): container finished" podID="02261047-2e53-4b34-bde5-5da7facb243c" containerID="8a9cc594f867e17e58e08d3984767170690347ae2f35b2e53796aa7c6d0db5f2" exitCode=0 Feb 02 15:38:41 crc kubenswrapper[4733]: I0202 15:38:41.703752 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/openstack-galera-2" event={"ID":"02261047-2e53-4b34-bde5-5da7facb243c","Type":"ContainerDied","Data":"8a9cc594f867e17e58e08d3984767170690347ae2f35b2e53796aa7c6d0db5f2"} Feb 02 15:38:41 crc kubenswrapper[4733]: I0202 15:38:41.706005 4733 generic.go:334] "Generic (PLEG): container finished" podID="79f9eeab-dccd-4404-8179-fc3f776d8b22" containerID="dcbd76747565b8c286db12738d84d8bb69ce0b89f08b555bfeb8f3b50e47baf1" exitCode=1 Feb 02 15:38:41 crc kubenswrapper[4733]: I0202 15:38:41.706057 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/keystonec686-account-delete-zqss2" event={"ID":"79f9eeab-dccd-4404-8179-fc3f776d8b22","Type":"ContainerDied","Data":"dcbd76747565b8c286db12738d84d8bb69ce0b89f08b555bfeb8f3b50e47baf1"} Feb 02 15:38:41 crc kubenswrapper[4733]: I0202 15:38:41.706096 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/keystonec686-account-delete-zqss2" event={"ID":"79f9eeab-dccd-4404-8179-fc3f776d8b22","Type":"ContainerStarted","Data":"81c30aa929390953e710af0da8ec697184872b5c3b9ac749b42d4051e0b2ad0b"} Feb 02 15:38:41 crc kubenswrapper[4733]: I0202 15:38:41.706656 4733 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="manila-kuttl-tests/keystonec686-account-delete-zqss2" secret="" err="secret \"galera-openstack-dockercfg-4hmlb\" not found" Feb 02 15:38:41 crc kubenswrapper[4733]: I0202 15:38:41.706710 4733 scope.go:117] "RemoveContainer" containerID="dcbd76747565b8c286db12738d84d8bb69ce0b89f08b555bfeb8f3b50e47baf1" Feb 02 15:38:41 crc kubenswrapper[4733]: E0202 15:38:41.809023 4733 configmap.go:193] Couldn't get configMap manila-kuttl-tests/openstack-scripts: configmap "openstack-scripts" not found Feb 02 15:38:41 crc kubenswrapper[4733]: E0202 15:38:41.809117 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/79f9eeab-dccd-4404-8179-fc3f776d8b22-operator-scripts podName:79f9eeab-dccd-4404-8179-fc3f776d8b22 nodeName:}" failed. No retries permitted until 2026-02-02 15:38:42.309096117 +0000 UTC m=+1405.760557485 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/79f9eeab-dccd-4404-8179-fc3f776d8b22-operator-scripts") pod "keystonec686-account-delete-zqss2" (UID: "79f9eeab-dccd-4404-8179-fc3f776d8b22") : configmap "openstack-scripts" not found Feb 02 15:38:41 crc kubenswrapper[4733]: I0202 15:38:41.827544 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["manila-kuttl-tests/rabbitmq-server-0"] Feb 02 15:38:41 crc kubenswrapper[4733]: I0202 15:38:41.915456 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/openstack-galera-2" Feb 02 15:38:42 crc kubenswrapper[4733]: I0202 15:38:42.010790 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/02261047-2e53-4b34-bde5-5da7facb243c-kolla-config\") pod \"02261047-2e53-4b34-bde5-5da7facb243c\" (UID: \"02261047-2e53-4b34-bde5-5da7facb243c\") " Feb 02 15:38:42 crc kubenswrapper[4733]: I0202 15:38:42.010873 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"02261047-2e53-4b34-bde5-5da7facb243c\" (UID: \"02261047-2e53-4b34-bde5-5da7facb243c\") " Feb 02 15:38:42 crc kubenswrapper[4733]: I0202 15:38:42.010920 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6q42z\" (UniqueName: \"kubernetes.io/projected/02261047-2e53-4b34-bde5-5da7facb243c-kube-api-access-6q42z\") pod \"02261047-2e53-4b34-bde5-5da7facb243c\" (UID: \"02261047-2e53-4b34-bde5-5da7facb243c\") " Feb 02 15:38:42 crc kubenswrapper[4733]: I0202 15:38:42.010964 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/02261047-2e53-4b34-bde5-5da7facb243c-config-data-generated\") pod \"02261047-2e53-4b34-bde5-5da7facb243c\" (UID: \"02261047-2e53-4b34-bde5-5da7facb243c\") " Feb 02 15:38:42 crc kubenswrapper[4733]: I0202 15:38:42.011028 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/02261047-2e53-4b34-bde5-5da7facb243c-config-data-default\") pod \"02261047-2e53-4b34-bde5-5da7facb243c\" (UID: \"02261047-2e53-4b34-bde5-5da7facb243c\") " Feb 02 15:38:42 crc kubenswrapper[4733]: I0202 15:38:42.011224 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02261047-2e53-4b34-bde5-5da7facb243c-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "02261047-2e53-4b34-bde5-5da7facb243c" (UID: "02261047-2e53-4b34-bde5-5da7facb243c"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:38:42 crc kubenswrapper[4733]: I0202 15:38:42.011524 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/02261047-2e53-4b34-bde5-5da7facb243c-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "02261047-2e53-4b34-bde5-5da7facb243c" (UID: "02261047-2e53-4b34-bde5-5da7facb243c"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:38:42 crc kubenswrapper[4733]: I0202 15:38:42.011629 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02261047-2e53-4b34-bde5-5da7facb243c-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "02261047-2e53-4b34-bde5-5da7facb243c" (UID: "02261047-2e53-4b34-bde5-5da7facb243c"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:38:42 crc kubenswrapper[4733]: I0202 15:38:42.011701 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/02261047-2e53-4b34-bde5-5da7facb243c-operator-scripts\") pod \"02261047-2e53-4b34-bde5-5da7facb243c\" (UID: \"02261047-2e53-4b34-bde5-5da7facb243c\") " Feb 02 15:38:42 crc kubenswrapper[4733]: I0202 15:38:42.012063 4733 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/02261047-2e53-4b34-bde5-5da7facb243c-kolla-config\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:42 crc kubenswrapper[4733]: I0202 15:38:42.012085 4733 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/02261047-2e53-4b34-bde5-5da7facb243c-config-data-generated\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:42 crc kubenswrapper[4733]: I0202 15:38:42.012099 4733 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/02261047-2e53-4b34-bde5-5da7facb243c-config-data-default\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:42 crc kubenswrapper[4733]: I0202 15:38:42.012186 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02261047-2e53-4b34-bde5-5da7facb243c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "02261047-2e53-4b34-bde5-5da7facb243c" (UID: "02261047-2e53-4b34-bde5-5da7facb243c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:38:42 crc kubenswrapper[4733]: I0202 15:38:42.017560 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02261047-2e53-4b34-bde5-5da7facb243c-kube-api-access-6q42z" (OuterVolumeSpecName: "kube-api-access-6q42z") pod "02261047-2e53-4b34-bde5-5da7facb243c" (UID: "02261047-2e53-4b34-bde5-5da7facb243c"). InnerVolumeSpecName "kube-api-access-6q42z". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:38:42 crc kubenswrapper[4733]: I0202 15:38:42.022126 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "mysql-db") pod "02261047-2e53-4b34-bde5-5da7facb243c" (UID: "02261047-2e53-4b34-bde5-5da7facb243c"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 02 15:38:42 crc kubenswrapper[4733]: I0202 15:38:42.112904 4733 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Feb 02 15:38:42 crc kubenswrapper[4733]: I0202 15:38:42.112933 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6q42z\" (UniqueName: \"kubernetes.io/projected/02261047-2e53-4b34-bde5-5da7facb243c-kube-api-access-6q42z\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:42 crc kubenswrapper[4733]: I0202 15:38:42.112943 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/02261047-2e53-4b34-bde5-5da7facb243c-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:42 crc kubenswrapper[4733]: I0202 15:38:42.126914 4733 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Feb 02 15:38:42 crc kubenswrapper[4733]: I0202 15:38:42.218233 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/rabbitmq-server-0"] Feb 02 15:38:42 crc kubenswrapper[4733]: I0202 15:38:42.218932 4733 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:42 crc kubenswrapper[4733]: E0202 15:38:42.320276 4733 configmap.go:193] Couldn't get configMap manila-kuttl-tests/openstack-scripts: configmap "openstack-scripts" not found Feb 02 15:38:42 crc kubenswrapper[4733]: E0202 15:38:42.320344 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/79f9eeab-dccd-4404-8179-fc3f776d8b22-operator-scripts podName:79f9eeab-dccd-4404-8179-fc3f776d8b22 nodeName:}" failed. No retries permitted until 2026-02-02 15:38:43.320330242 +0000 UTC m=+1406.771791600 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/79f9eeab-dccd-4404-8179-fc3f776d8b22-operator-scripts") pod "keystonec686-account-delete-zqss2" (UID: "79f9eeab-dccd-4404-8179-fc3f776d8b22") : configmap "openstack-scripts" not found Feb 02 15:38:42 crc kubenswrapper[4733]: I0202 15:38:42.733268 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/openstack-galera-2" event={"ID":"02261047-2e53-4b34-bde5-5da7facb243c","Type":"ContainerDied","Data":"31ac1ca7ba19035b1b7157aafd05dadb7f92843c69e455f557bd46e0662d7cdf"} Feb 02 15:38:42 crc kubenswrapper[4733]: I0202 15:38:42.733640 4733 scope.go:117] "RemoveContainer" containerID="8a9cc594f867e17e58e08d3984767170690347ae2f35b2e53796aa7c6d0db5f2" Feb 02 15:38:42 crc kubenswrapper[4733]: I0202 15:38:42.733535 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/openstack-galera-2" Feb 02 15:38:42 crc kubenswrapper[4733]: I0202 15:38:42.735549 4733 generic.go:334] "Generic (PLEG): container finished" podID="e8f2d3a6-ad9f-4e1c-8f9a-41f8c7efcdac" containerID="60f342505fd409f05a577cf0a9883bd43822c387690a1b4617f376e8acec2dc4" exitCode=0 Feb 02 15:38:42 crc kubenswrapper[4733]: I0202 15:38:42.735638 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/memcached-0" event={"ID":"e8f2d3a6-ad9f-4e1c-8f9a-41f8c7efcdac","Type":"ContainerDied","Data":"60f342505fd409f05a577cf0a9883bd43822c387690a1b4617f376e8acec2dc4"} Feb 02 15:38:42 crc kubenswrapper[4733]: I0202 15:38:42.738021 4733 generic.go:334] "Generic (PLEG): container finished" podID="79f9eeab-dccd-4404-8179-fc3f776d8b22" containerID="5a580dc0dc11b40f3390a7c1a378ba8805983a64d2d3cedbeaf23f4a821f2126" exitCode=1 Feb 02 15:38:42 crc kubenswrapper[4733]: I0202 15:38:42.738851 4733 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="manila-kuttl-tests/keystonec686-account-delete-zqss2" secret="" err="secret \"galera-openstack-dockercfg-4hmlb\" not found" Feb 02 15:38:42 crc kubenswrapper[4733]: I0202 15:38:42.738886 4733 scope.go:117] "RemoveContainer" containerID="5a580dc0dc11b40f3390a7c1a378ba8805983a64d2d3cedbeaf23f4a821f2126" Feb 02 15:38:42 crc kubenswrapper[4733]: E0202 15:38:42.739094 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-delete\" with CrashLoopBackOff: \"back-off 10s restarting failed container=mariadb-account-delete pod=keystonec686-account-delete-zqss2_manila-kuttl-tests(79f9eeab-dccd-4404-8179-fc3f776d8b22)\"" pod="manila-kuttl-tests/keystonec686-account-delete-zqss2" podUID="79f9eeab-dccd-4404-8179-fc3f776d8b22" Feb 02 15:38:42 crc kubenswrapper[4733]: I0202 15:38:42.739124 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/keystonec686-account-delete-zqss2" event={"ID":"79f9eeab-dccd-4404-8179-fc3f776d8b22","Type":"ContainerDied","Data":"5a580dc0dc11b40f3390a7c1a378ba8805983a64d2d3cedbeaf23f4a821f2126"} Feb 02 15:38:42 crc kubenswrapper[4733]: I0202 15:38:42.783329 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/ceph"] Feb 02 15:38:42 crc kubenswrapper[4733]: I0202 15:38:42.783626 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="manila-kuttl-tests/ceph" podUID="c71eca7f-f211-40bc-9116-f3c246c46e54" containerName="ceph" containerID="cri-o://00f13c6c3710736a5fd10709601532b0ab4f0709341f931f4399424374017bce" gracePeriod=30 Feb 02 15:38:42 crc kubenswrapper[4733]: I0202 15:38:42.787928 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/openstack-galera-2"] Feb 02 15:38:42 crc kubenswrapper[4733]: I0202 15:38:42.788826 4733 scope.go:117] "RemoveContainer" containerID="ca13066326a1d72613f501d942b75303c019bac4f9000af669ea09823a2af28b" Feb 02 15:38:42 crc kubenswrapper[4733]: I0202 15:38:42.791923 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["manila-kuttl-tests/openstack-galera-2"] Feb 02 15:38:42 crc kubenswrapper[4733]: I0202 15:38:42.815194 4733 scope.go:117] "RemoveContainer" containerID="dcbd76747565b8c286db12738d84d8bb69ce0b89f08b555bfeb8f3b50e47baf1" Feb 02 15:38:42 crc kubenswrapper[4733]: I0202 15:38:42.816707 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="manila-kuttl-tests/rabbitmq-server-0" podUID="f45e90dd-ab9d-4c27-8449-98cf52f48c87" containerName="rabbitmq" containerID="cri-o://40d292ee4fa8872456f484269b6ff9ad31d0092c116a9313314a8f34ed13addd" gracePeriod=604800 Feb 02 15:38:42 crc kubenswrapper[4733]: I0202 15:38:42.885312 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/memcached-0" Feb 02 15:38:42 crc kubenswrapper[4733]: I0202 15:38:42.976342 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="manila-kuttl-tests/openstack-galera-1" podUID="d63f3e93-f45f-4f7d-a29b-2eaae536fa5d" containerName="galera" containerID="cri-o://77d577f4756bae43064e83c2d18077bad1b459a16e5122f3a5125f7ac44f7514" gracePeriod=28 Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.031637 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zj2tt\" (UniqueName: \"kubernetes.io/projected/e8f2d3a6-ad9f-4e1c-8f9a-41f8c7efcdac-kube-api-access-zj2tt\") pod \"e8f2d3a6-ad9f-4e1c-8f9a-41f8c7efcdac\" (UID: \"e8f2d3a6-ad9f-4e1c-8f9a-41f8c7efcdac\") " Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.031703 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e8f2d3a6-ad9f-4e1c-8f9a-41f8c7efcdac-config-data\") pod \"e8f2d3a6-ad9f-4e1c-8f9a-41f8c7efcdac\" (UID: \"e8f2d3a6-ad9f-4e1c-8f9a-41f8c7efcdac\") " Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.031774 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e8f2d3a6-ad9f-4e1c-8f9a-41f8c7efcdac-kolla-config\") pod \"e8f2d3a6-ad9f-4e1c-8f9a-41f8c7efcdac\" (UID: \"e8f2d3a6-ad9f-4e1c-8f9a-41f8c7efcdac\") " Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.032414 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8f2d3a6-ad9f-4e1c-8f9a-41f8c7efcdac-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "e8f2d3a6-ad9f-4e1c-8f9a-41f8c7efcdac" (UID: "e8f2d3a6-ad9f-4e1c-8f9a-41f8c7efcdac"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.032439 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8f2d3a6-ad9f-4e1c-8f9a-41f8c7efcdac-config-data" (OuterVolumeSpecName: "config-data") pod "e8f2d3a6-ad9f-4e1c-8f9a-41f8c7efcdac" (UID: "e8f2d3a6-ad9f-4e1c-8f9a-41f8c7efcdac"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.036934 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8f2d3a6-ad9f-4e1c-8f9a-41f8c7efcdac-kube-api-access-zj2tt" (OuterVolumeSpecName: "kube-api-access-zj2tt") pod "e8f2d3a6-ad9f-4e1c-8f9a-41f8c7efcdac" (UID: "e8f2d3a6-ad9f-4e1c-8f9a-41f8c7efcdac"). InnerVolumeSpecName "kube-api-access-zj2tt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.133793 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zj2tt\" (UniqueName: \"kubernetes.io/projected/e8f2d3a6-ad9f-4e1c-8f9a-41f8c7efcdac-kube-api-access-zj2tt\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.133847 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e8f2d3a6-ad9f-4e1c-8f9a-41f8c7efcdac-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.133866 4733 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e8f2d3a6-ad9f-4e1c-8f9a-41f8c7efcdac-kolla-config\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.268207 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02261047-2e53-4b34-bde5-5da7facb243c" path="/var/lib/kubelet/pods/02261047-2e53-4b34-bde5-5da7facb243c/volumes" Feb 02 15:38:43 crc kubenswrapper[4733]: E0202 15:38:43.336699 4733 configmap.go:193] Couldn't get configMap manila-kuttl-tests/openstack-scripts: configmap "openstack-scripts" not found Feb 02 15:38:43 crc kubenswrapper[4733]: E0202 15:38:43.337127 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/79f9eeab-dccd-4404-8179-fc3f776d8b22-operator-scripts podName:79f9eeab-dccd-4404-8179-fc3f776d8b22 nodeName:}" failed. No retries permitted until 2026-02-02 15:38:45.337110162 +0000 UTC m=+1408.788571520 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/79f9eeab-dccd-4404-8179-fc3f776d8b22-operator-scripts") pod "keystonec686-account-delete-zqss2" (UID: "79f9eeab-dccd-4404-8179-fc3f776d8b22") : configmap "openstack-scripts" not found Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.680521 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/keystone-5f5c86bdc8-jpphg" Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.745752 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a972e8cd-cae8-4894-b2d3-7bfff75acc1c-scripts\") pod \"a972e8cd-cae8-4894-b2d3-7bfff75acc1c\" (UID: \"a972e8cd-cae8-4894-b2d3-7bfff75acc1c\") " Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.745821 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a972e8cd-cae8-4894-b2d3-7bfff75acc1c-config-data\") pod \"a972e8cd-cae8-4894-b2d3-7bfff75acc1c\" (UID: \"a972e8cd-cae8-4894-b2d3-7bfff75acc1c\") " Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.745845 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a972e8cd-cae8-4894-b2d3-7bfff75acc1c-credential-keys\") pod \"a972e8cd-cae8-4894-b2d3-7bfff75acc1c\" (UID: \"a972e8cd-cae8-4894-b2d3-7bfff75acc1c\") " Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.745896 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a972e8cd-cae8-4894-b2d3-7bfff75acc1c-fernet-keys\") pod \"a972e8cd-cae8-4894-b2d3-7bfff75acc1c\" (UID: \"a972e8cd-cae8-4894-b2d3-7bfff75acc1c\") " Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.745951 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5nr7m\" (UniqueName: \"kubernetes.io/projected/a972e8cd-cae8-4894-b2d3-7bfff75acc1c-kube-api-access-5nr7m\") pod \"a972e8cd-cae8-4894-b2d3-7bfff75acc1c\" (UID: \"a972e8cd-cae8-4894-b2d3-7bfff75acc1c\") " Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.748070 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9gtt4"] Feb 02 15:38:43 crc kubenswrapper[4733]: E0202 15:38:43.748463 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02261047-2e53-4b34-bde5-5da7facb243c" containerName="galera" Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.748485 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="02261047-2e53-4b34-bde5-5da7facb243c" containerName="galera" Feb 02 15:38:43 crc kubenswrapper[4733]: E0202 15:38:43.748504 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8f2d3a6-ad9f-4e1c-8f9a-41f8c7efcdac" containerName="memcached" Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.748516 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8f2d3a6-ad9f-4e1c-8f9a-41f8c7efcdac" containerName="memcached" Feb 02 15:38:43 crc kubenswrapper[4733]: E0202 15:38:43.748535 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a972e8cd-cae8-4894-b2d3-7bfff75acc1c" containerName="keystone-api" Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.748546 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="a972e8cd-cae8-4894-b2d3-7bfff75acc1c" containerName="keystone-api" Feb 02 15:38:43 crc kubenswrapper[4733]: E0202 15:38:43.748569 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02261047-2e53-4b34-bde5-5da7facb243c" containerName="mysql-bootstrap" Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.748578 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="02261047-2e53-4b34-bde5-5da7facb243c" containerName="mysql-bootstrap" Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.748750 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="02261047-2e53-4b34-bde5-5da7facb243c" containerName="galera" Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.748775 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="a972e8cd-cae8-4894-b2d3-7bfff75acc1c" containerName="keystone-api" Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.748797 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8f2d3a6-ad9f-4e1c-8f9a-41f8c7efcdac" containerName="memcached" Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.750145 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9gtt4" Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.751441 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a972e8cd-cae8-4894-b2d3-7bfff75acc1c-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "a972e8cd-cae8-4894-b2d3-7bfff75acc1c" (UID: "a972e8cd-cae8-4894-b2d3-7bfff75acc1c"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.751828 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a972e8cd-cae8-4894-b2d3-7bfff75acc1c-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "a972e8cd-cae8-4894-b2d3-7bfff75acc1c" (UID: "a972e8cd-cae8-4894-b2d3-7bfff75acc1c"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.752513 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a972e8cd-cae8-4894-b2d3-7bfff75acc1c-kube-api-access-5nr7m" (OuterVolumeSpecName: "kube-api-access-5nr7m") pod "a972e8cd-cae8-4894-b2d3-7bfff75acc1c" (UID: "a972e8cd-cae8-4894-b2d3-7bfff75acc1c"). InnerVolumeSpecName "kube-api-access-5nr7m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.754341 4733 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="manila-kuttl-tests/keystonec686-account-delete-zqss2" secret="" err="secret \"galera-openstack-dockercfg-4hmlb\" not found" Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.754386 4733 scope.go:117] "RemoveContainer" containerID="5a580dc0dc11b40f3390a7c1a378ba8805983a64d2d3cedbeaf23f4a821f2126" Feb 02 15:38:43 crc kubenswrapper[4733]: E0202 15:38:43.754772 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-delete\" with CrashLoopBackOff: \"back-off 10s restarting failed container=mariadb-account-delete pod=keystonec686-account-delete-zqss2_manila-kuttl-tests(79f9eeab-dccd-4404-8179-fc3f776d8b22)\"" pod="manila-kuttl-tests/keystonec686-account-delete-zqss2" podUID="79f9eeab-dccd-4404-8179-fc3f776d8b22" Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.758516 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a972e8cd-cae8-4894-b2d3-7bfff75acc1c-scripts" (OuterVolumeSpecName: "scripts") pod "a972e8cd-cae8-4894-b2d3-7bfff75acc1c" (UID: "a972e8cd-cae8-4894-b2d3-7bfff75acc1c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.765906 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9gtt4"] Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.778190 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a972e8cd-cae8-4894-b2d3-7bfff75acc1c-config-data" (OuterVolumeSpecName: "config-data") pod "a972e8cd-cae8-4894-b2d3-7bfff75acc1c" (UID: "a972e8cd-cae8-4894-b2d3-7bfff75acc1c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.780525 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/memcached-0" event={"ID":"e8f2d3a6-ad9f-4e1c-8f9a-41f8c7efcdac","Type":"ContainerDied","Data":"4f75aaa4c7a7096a3bad8398b01deb8d12d4e519d6c78295f64cd3f241439ba6"} Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.780577 4733 scope.go:117] "RemoveContainer" containerID="60f342505fd409f05a577cf0a9883bd43822c387690a1b4617f376e8acec2dc4" Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.780679 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/memcached-0" Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.788248 4733 generic.go:334] "Generic (PLEG): container finished" podID="a972e8cd-cae8-4894-b2d3-7bfff75acc1c" containerID="818ce15b55c28e765033c326ced7cc8bb1dd7b47fe329b86dc8b9830026fccea" exitCode=0 Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.788297 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/keystone-5f5c86bdc8-jpphg" event={"ID":"a972e8cd-cae8-4894-b2d3-7bfff75acc1c","Type":"ContainerDied","Data":"818ce15b55c28e765033c326ced7cc8bb1dd7b47fe329b86dc8b9830026fccea"} Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.788705 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/keystone-5f5c86bdc8-jpphg" event={"ID":"a972e8cd-cae8-4894-b2d3-7bfff75acc1c","Type":"ContainerDied","Data":"5da8dbd95ec2753a5a0ea2dcb28b8f2288a6cac18e606e5f795f27e228c125f0"} Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.792969 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/keystone-5f5c86bdc8-jpphg" Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.816959 4733 scope.go:117] "RemoveContainer" containerID="818ce15b55c28e765033c326ced7cc8bb1dd7b47fe329b86dc8b9830026fccea" Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.833557 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/memcached-0"] Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.841333 4733 scope.go:117] "RemoveContainer" containerID="818ce15b55c28e765033c326ced7cc8bb1dd7b47fe329b86dc8b9830026fccea" Feb 02 15:38:43 crc kubenswrapper[4733]: E0202 15:38:43.842705 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"818ce15b55c28e765033c326ced7cc8bb1dd7b47fe329b86dc8b9830026fccea\": container with ID starting with 818ce15b55c28e765033c326ced7cc8bb1dd7b47fe329b86dc8b9830026fccea not found: ID does not exist" containerID="818ce15b55c28e765033c326ced7cc8bb1dd7b47fe329b86dc8b9830026fccea" Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.842732 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"818ce15b55c28e765033c326ced7cc8bb1dd7b47fe329b86dc8b9830026fccea"} err="failed to get container status \"818ce15b55c28e765033c326ced7cc8bb1dd7b47fe329b86dc8b9830026fccea\": rpc error: code = NotFound desc = could not find container \"818ce15b55c28e765033c326ced7cc8bb1dd7b47fe329b86dc8b9830026fccea\": container with ID starting with 818ce15b55c28e765033c326ced7cc8bb1dd7b47fe329b86dc8b9830026fccea not found: ID does not exist" Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.849468 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/484de6bc-40c4-4159-a4f5-9cf345357b40-catalog-content\") pod \"community-operators-9gtt4\" (UID: \"484de6bc-40c4-4159-a4f5-9cf345357b40\") " pod="openshift-marketplace/community-operators-9gtt4" Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.849567 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwwkk\" (UniqueName: \"kubernetes.io/projected/484de6bc-40c4-4159-a4f5-9cf345357b40-kube-api-access-qwwkk\") pod \"community-operators-9gtt4\" (UID: \"484de6bc-40c4-4159-a4f5-9cf345357b40\") " pod="openshift-marketplace/community-operators-9gtt4" Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.849597 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/484de6bc-40c4-4159-a4f5-9cf345357b40-utilities\") pod \"community-operators-9gtt4\" (UID: \"484de6bc-40c4-4159-a4f5-9cf345357b40\") " pod="openshift-marketplace/community-operators-9gtt4" Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.849656 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a972e8cd-cae8-4894-b2d3-7bfff75acc1c-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.849668 4733 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a972e8cd-cae8-4894-b2d3-7bfff75acc1c-credential-keys\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.849696 4733 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a972e8cd-cae8-4894-b2d3-7bfff75acc1c-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.849706 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5nr7m\" (UniqueName: \"kubernetes.io/projected/a972e8cd-cae8-4894-b2d3-7bfff75acc1c-kube-api-access-5nr7m\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.849714 4733 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a972e8cd-cae8-4894-b2d3-7bfff75acc1c-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.864614 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["manila-kuttl-tests/memcached-0"] Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.871551 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/keystone-5f5c86bdc8-jpphg"] Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.874890 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["manila-kuttl-tests/keystone-5f5c86bdc8-jpphg"] Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.950666 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/484de6bc-40c4-4159-a4f5-9cf345357b40-catalog-content\") pod \"community-operators-9gtt4\" (UID: \"484de6bc-40c4-4159-a4f5-9cf345357b40\") " pod="openshift-marketplace/community-operators-9gtt4" Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.950745 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwwkk\" (UniqueName: \"kubernetes.io/projected/484de6bc-40c4-4159-a4f5-9cf345357b40-kube-api-access-qwwkk\") pod \"community-operators-9gtt4\" (UID: \"484de6bc-40c4-4159-a4f5-9cf345357b40\") " pod="openshift-marketplace/community-operators-9gtt4" Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.950776 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/484de6bc-40c4-4159-a4f5-9cf345357b40-utilities\") pod \"community-operators-9gtt4\" (UID: \"484de6bc-40c4-4159-a4f5-9cf345357b40\") " pod="openshift-marketplace/community-operators-9gtt4" Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.951392 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/484de6bc-40c4-4159-a4f5-9cf345357b40-utilities\") pod \"community-operators-9gtt4\" (UID: \"484de6bc-40c4-4159-a4f5-9cf345357b40\") " pod="openshift-marketplace/community-operators-9gtt4" Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.951588 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/484de6bc-40c4-4159-a4f5-9cf345357b40-catalog-content\") pod \"community-operators-9gtt4\" (UID: \"484de6bc-40c4-4159-a4f5-9cf345357b40\") " pod="openshift-marketplace/community-operators-9gtt4" Feb 02 15:38:43 crc kubenswrapper[4733]: I0202 15:38:43.969848 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwwkk\" (UniqueName: \"kubernetes.io/projected/484de6bc-40c4-4159-a4f5-9cf345357b40-kube-api-access-qwwkk\") pod \"community-operators-9gtt4\" (UID: \"484de6bc-40c4-4159-a4f5-9cf345357b40\") " pod="openshift-marketplace/community-operators-9gtt4" Feb 02 15:38:44 crc kubenswrapper[4733]: I0202 15:38:44.123263 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9gtt4" Feb 02 15:38:44 crc kubenswrapper[4733]: I0202 15:38:44.230541 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/rabbitmq-server-0" Feb 02 15:38:44 crc kubenswrapper[4733]: I0202 15:38:44.356536 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f45e90dd-ab9d-4c27-8449-98cf52f48c87-rabbitmq-erlang-cookie\") pod \"f45e90dd-ab9d-4c27-8449-98cf52f48c87\" (UID: \"f45e90dd-ab9d-4c27-8449-98cf52f48c87\") " Feb 02 15:38:44 crc kubenswrapper[4733]: I0202 15:38:44.356823 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f45e90dd-ab9d-4c27-8449-98cf52f48c87-rabbitmq-plugins\") pod \"f45e90dd-ab9d-4c27-8449-98cf52f48c87\" (UID: \"f45e90dd-ab9d-4c27-8449-98cf52f48c87\") " Feb 02 15:38:44 crc kubenswrapper[4733]: I0202 15:38:44.356844 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f45e90dd-ab9d-4c27-8449-98cf52f48c87-plugins-conf\") pod \"f45e90dd-ab9d-4c27-8449-98cf52f48c87\" (UID: \"f45e90dd-ab9d-4c27-8449-98cf52f48c87\") " Feb 02 15:38:44 crc kubenswrapper[4733]: I0202 15:38:44.356908 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f45e90dd-ab9d-4c27-8449-98cf52f48c87-rabbitmq-confd\") pod \"f45e90dd-ab9d-4c27-8449-98cf52f48c87\" (UID: \"f45e90dd-ab9d-4c27-8449-98cf52f48c87\") " Feb 02 15:38:44 crc kubenswrapper[4733]: I0202 15:38:44.357037 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cee723a2-19bb-4ddf-b89c-19eee3872720\") pod \"f45e90dd-ab9d-4c27-8449-98cf52f48c87\" (UID: \"f45e90dd-ab9d-4c27-8449-98cf52f48c87\") " Feb 02 15:38:44 crc kubenswrapper[4733]: I0202 15:38:44.357095 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mfh6r\" (UniqueName: \"kubernetes.io/projected/f45e90dd-ab9d-4c27-8449-98cf52f48c87-kube-api-access-mfh6r\") pod \"f45e90dd-ab9d-4c27-8449-98cf52f48c87\" (UID: \"f45e90dd-ab9d-4c27-8449-98cf52f48c87\") " Feb 02 15:38:44 crc kubenswrapper[4733]: I0202 15:38:44.357143 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f45e90dd-ab9d-4c27-8449-98cf52f48c87-pod-info\") pod \"f45e90dd-ab9d-4c27-8449-98cf52f48c87\" (UID: \"f45e90dd-ab9d-4c27-8449-98cf52f48c87\") " Feb 02 15:38:44 crc kubenswrapper[4733]: I0202 15:38:44.357195 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f45e90dd-ab9d-4c27-8449-98cf52f48c87-erlang-cookie-secret\") pod \"f45e90dd-ab9d-4c27-8449-98cf52f48c87\" (UID: \"f45e90dd-ab9d-4c27-8449-98cf52f48c87\") " Feb 02 15:38:44 crc kubenswrapper[4733]: I0202 15:38:44.357264 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f45e90dd-ab9d-4c27-8449-98cf52f48c87-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "f45e90dd-ab9d-4c27-8449-98cf52f48c87" (UID: "f45e90dd-ab9d-4c27-8449-98cf52f48c87"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:38:44 crc kubenswrapper[4733]: I0202 15:38:44.357282 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f45e90dd-ab9d-4c27-8449-98cf52f48c87-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "f45e90dd-ab9d-4c27-8449-98cf52f48c87" (UID: "f45e90dd-ab9d-4c27-8449-98cf52f48c87"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:38:44 crc kubenswrapper[4733]: I0202 15:38:44.357311 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f45e90dd-ab9d-4c27-8449-98cf52f48c87-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "f45e90dd-ab9d-4c27-8449-98cf52f48c87" (UID: "f45e90dd-ab9d-4c27-8449-98cf52f48c87"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:38:44 crc kubenswrapper[4733]: I0202 15:38:44.357630 4733 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f45e90dd-ab9d-4c27-8449-98cf52f48c87-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:44 crc kubenswrapper[4733]: I0202 15:38:44.357653 4733 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f45e90dd-ab9d-4c27-8449-98cf52f48c87-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:44 crc kubenswrapper[4733]: I0202 15:38:44.357666 4733 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f45e90dd-ab9d-4c27-8449-98cf52f48c87-plugins-conf\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:44 crc kubenswrapper[4733]: I0202 15:38:44.360444 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/f45e90dd-ab9d-4c27-8449-98cf52f48c87-pod-info" (OuterVolumeSpecName: "pod-info") pod "f45e90dd-ab9d-4c27-8449-98cf52f48c87" (UID: "f45e90dd-ab9d-4c27-8449-98cf52f48c87"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Feb 02 15:38:44 crc kubenswrapper[4733]: I0202 15:38:44.360792 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f45e90dd-ab9d-4c27-8449-98cf52f48c87-kube-api-access-mfh6r" (OuterVolumeSpecName: "kube-api-access-mfh6r") pod "f45e90dd-ab9d-4c27-8449-98cf52f48c87" (UID: "f45e90dd-ab9d-4c27-8449-98cf52f48c87"). InnerVolumeSpecName "kube-api-access-mfh6r". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:38:44 crc kubenswrapper[4733]: I0202 15:38:44.364254 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f45e90dd-ab9d-4c27-8449-98cf52f48c87-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "f45e90dd-ab9d-4c27-8449-98cf52f48c87" (UID: "f45e90dd-ab9d-4c27-8449-98cf52f48c87"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:38:44 crc kubenswrapper[4733]: I0202 15:38:44.371681 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cee723a2-19bb-4ddf-b89c-19eee3872720" (OuterVolumeSpecName: "persistence") pod "f45e90dd-ab9d-4c27-8449-98cf52f48c87" (UID: "f45e90dd-ab9d-4c27-8449-98cf52f48c87"). InnerVolumeSpecName "pvc-cee723a2-19bb-4ddf-b89c-19eee3872720". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 02 15:38:44 crc kubenswrapper[4733]: I0202 15:38:44.423333 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f45e90dd-ab9d-4c27-8449-98cf52f48c87-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "f45e90dd-ab9d-4c27-8449-98cf52f48c87" (UID: "f45e90dd-ab9d-4c27-8449-98cf52f48c87"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:38:44 crc kubenswrapper[4733]: I0202 15:38:44.459146 4733 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f45e90dd-ab9d-4c27-8449-98cf52f48c87-pod-info\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:44 crc kubenswrapper[4733]: I0202 15:38:44.459186 4733 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f45e90dd-ab9d-4c27-8449-98cf52f48c87-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:44 crc kubenswrapper[4733]: I0202 15:38:44.459200 4733 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f45e90dd-ab9d-4c27-8449-98cf52f48c87-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:44 crc kubenswrapper[4733]: I0202 15:38:44.459230 4733 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-cee723a2-19bb-4ddf-b89c-19eee3872720\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cee723a2-19bb-4ddf-b89c-19eee3872720\") on node \"crc\" " Feb 02 15:38:44 crc kubenswrapper[4733]: I0202 15:38:44.459240 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mfh6r\" (UniqueName: \"kubernetes.io/projected/f45e90dd-ab9d-4c27-8449-98cf52f48c87-kube-api-access-mfh6r\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:44 crc kubenswrapper[4733]: I0202 15:38:44.471078 4733 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Feb 02 15:38:44 crc kubenswrapper[4733]: I0202 15:38:44.471229 4733 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-cee723a2-19bb-4ddf-b89c-19eee3872720" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cee723a2-19bb-4ddf-b89c-19eee3872720") on node "crc" Feb 02 15:38:44 crc kubenswrapper[4733]: I0202 15:38:44.560457 4733 reconciler_common.go:293] "Volume detached for volume \"pvc-cee723a2-19bb-4ddf-b89c-19eee3872720\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cee723a2-19bb-4ddf-b89c-19eee3872720\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:44 crc kubenswrapper[4733]: I0202 15:38:44.671827 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9gtt4"] Feb 02 15:38:44 crc kubenswrapper[4733]: W0202 15:38:44.674734 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod484de6bc_40c4_4159_a4f5_9cf345357b40.slice/crio-18a6cbd09141c7a50c8db5f6e1be47576829bf56bccc2297ef239428b0767d08 WatchSource:0}: Error finding container 18a6cbd09141c7a50c8db5f6e1be47576829bf56bccc2297ef239428b0767d08: Status 404 returned error can't find the container with id 18a6cbd09141c7a50c8db5f6e1be47576829bf56bccc2297ef239428b0767d08 Feb 02 15:38:44 crc kubenswrapper[4733]: I0202 15:38:44.802355 4733 generic.go:334] "Generic (PLEG): container finished" podID="d63f3e93-f45f-4f7d-a29b-2eaae536fa5d" containerID="77d577f4756bae43064e83c2d18077bad1b459a16e5122f3a5125f7ac44f7514" exitCode=0 Feb 02 15:38:44 crc kubenswrapper[4733]: I0202 15:38:44.802445 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/openstack-galera-1" event={"ID":"d63f3e93-f45f-4f7d-a29b-2eaae536fa5d","Type":"ContainerDied","Data":"77d577f4756bae43064e83c2d18077bad1b459a16e5122f3a5125f7ac44f7514"} Feb 02 15:38:44 crc kubenswrapper[4733]: I0202 15:38:44.803980 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9gtt4" event={"ID":"484de6bc-40c4-4159-a4f5-9cf345357b40","Type":"ContainerStarted","Data":"18a6cbd09141c7a50c8db5f6e1be47576829bf56bccc2297ef239428b0767d08"} Feb 02 15:38:44 crc kubenswrapper[4733]: I0202 15:38:44.807877 4733 generic.go:334] "Generic (PLEG): container finished" podID="f45e90dd-ab9d-4c27-8449-98cf52f48c87" containerID="40d292ee4fa8872456f484269b6ff9ad31d0092c116a9313314a8f34ed13addd" exitCode=0 Feb 02 15:38:44 crc kubenswrapper[4733]: I0202 15:38:44.807909 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/rabbitmq-server-0" event={"ID":"f45e90dd-ab9d-4c27-8449-98cf52f48c87","Type":"ContainerDied","Data":"40d292ee4fa8872456f484269b6ff9ad31d0092c116a9313314a8f34ed13addd"} Feb 02 15:38:44 crc kubenswrapper[4733]: I0202 15:38:44.807925 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/rabbitmq-server-0" event={"ID":"f45e90dd-ab9d-4c27-8449-98cf52f48c87","Type":"ContainerDied","Data":"ca0247762a0b2747313840f74a70bbe642746790e8a786c7dbfdc7b707a1a2a7"} Feb 02 15:38:44 crc kubenswrapper[4733]: I0202 15:38:44.807927 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/rabbitmq-server-0" Feb 02 15:38:44 crc kubenswrapper[4733]: I0202 15:38:44.807941 4733 scope.go:117] "RemoveContainer" containerID="40d292ee4fa8872456f484269b6ff9ad31d0092c116a9313314a8f34ed13addd" Feb 02 15:38:44 crc kubenswrapper[4733]: I0202 15:38:44.892011 4733 scope.go:117] "RemoveContainer" containerID="9a9ed87542bec5386676f7a6416db7de1ec8fc04680e6dbf0f22482fb372172c" Feb 02 15:38:44 crc kubenswrapper[4733]: I0202 15:38:44.929654 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/rabbitmq-server-0"] Feb 02 15:38:44 crc kubenswrapper[4733]: I0202 15:38:44.935716 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["manila-kuttl-tests/rabbitmq-server-0"] Feb 02 15:38:44 crc kubenswrapper[4733]: I0202 15:38:44.939773 4733 scope.go:117] "RemoveContainer" containerID="40d292ee4fa8872456f484269b6ff9ad31d0092c116a9313314a8f34ed13addd" Feb 02 15:38:44 crc kubenswrapper[4733]: E0202 15:38:44.940278 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40d292ee4fa8872456f484269b6ff9ad31d0092c116a9313314a8f34ed13addd\": container with ID starting with 40d292ee4fa8872456f484269b6ff9ad31d0092c116a9313314a8f34ed13addd not found: ID does not exist" containerID="40d292ee4fa8872456f484269b6ff9ad31d0092c116a9313314a8f34ed13addd" Feb 02 15:38:44 crc kubenswrapper[4733]: I0202 15:38:44.940308 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40d292ee4fa8872456f484269b6ff9ad31d0092c116a9313314a8f34ed13addd"} err="failed to get container status \"40d292ee4fa8872456f484269b6ff9ad31d0092c116a9313314a8f34ed13addd\": rpc error: code = NotFound desc = could not find container \"40d292ee4fa8872456f484269b6ff9ad31d0092c116a9313314a8f34ed13addd\": container with ID starting with 40d292ee4fa8872456f484269b6ff9ad31d0092c116a9313314a8f34ed13addd not found: ID does not exist" Feb 02 15:38:44 crc kubenswrapper[4733]: I0202 15:38:44.940332 4733 scope.go:117] "RemoveContainer" containerID="9a9ed87542bec5386676f7a6416db7de1ec8fc04680e6dbf0f22482fb372172c" Feb 02 15:38:44 crc kubenswrapper[4733]: E0202 15:38:44.940653 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a9ed87542bec5386676f7a6416db7de1ec8fc04680e6dbf0f22482fb372172c\": container with ID starting with 9a9ed87542bec5386676f7a6416db7de1ec8fc04680e6dbf0f22482fb372172c not found: ID does not exist" containerID="9a9ed87542bec5386676f7a6416db7de1ec8fc04680e6dbf0f22482fb372172c" Feb 02 15:38:44 crc kubenswrapper[4733]: I0202 15:38:44.940685 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a9ed87542bec5386676f7a6416db7de1ec8fc04680e6dbf0f22482fb372172c"} err="failed to get container status \"9a9ed87542bec5386676f7a6416db7de1ec8fc04680e6dbf0f22482fb372172c\": rpc error: code = NotFound desc = could not find container \"9a9ed87542bec5386676f7a6416db7de1ec8fc04680e6dbf0f22482fb372172c\": container with ID starting with 9a9ed87542bec5386676f7a6416db7de1ec8fc04680e6dbf0f22482fb372172c not found: ID does not exist" Feb 02 15:38:44 crc kubenswrapper[4733]: I0202 15:38:44.999108 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="manila-kuttl-tests/openstack-galera-0" podUID="f177ebf9-728b-4ba2-b9b7-6c676f9d014a" containerName="galera" containerID="cri-o://69b15a1f6699ebd76df18a6d10489a9a8f470c2c56ccab52ad3c08a533bb0a92" gracePeriod=26 Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.075607 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/openstack-galera-1" Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.096210 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/keystone-db-create-s5qkh"] Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.106531 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["manila-kuttl-tests/keystone-db-create-s5qkh"] Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.132329 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/keystone-c686-account-create-update-msktt"] Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.148947 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/keystonec686-account-delete-zqss2"] Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.155975 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["manila-kuttl-tests/keystone-c686-account-create-update-msktt"] Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.168181 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"d63f3e93-f45f-4f7d-a29b-2eaae536fa5d\" (UID: \"d63f3e93-f45f-4f7d-a29b-2eaae536fa5d\") " Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.168232 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d63f3e93-f45f-4f7d-a29b-2eaae536fa5d-config-data-generated\") pod \"d63f3e93-f45f-4f7d-a29b-2eaae536fa5d\" (UID: \"d63f3e93-f45f-4f7d-a29b-2eaae536fa5d\") " Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.168268 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d63f3e93-f45f-4f7d-a29b-2eaae536fa5d-kolla-config\") pod \"d63f3e93-f45f-4f7d-a29b-2eaae536fa5d\" (UID: \"d63f3e93-f45f-4f7d-a29b-2eaae536fa5d\") " Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.168332 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d63f3e93-f45f-4f7d-a29b-2eaae536fa5d-config-data-default\") pod \"d63f3e93-f45f-4f7d-a29b-2eaae536fa5d\" (UID: \"d63f3e93-f45f-4f7d-a29b-2eaae536fa5d\") " Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.168386 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d63f3e93-f45f-4f7d-a29b-2eaae536fa5d-operator-scripts\") pod \"d63f3e93-f45f-4f7d-a29b-2eaae536fa5d\" (UID: \"d63f3e93-f45f-4f7d-a29b-2eaae536fa5d\") " Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.168403 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xsq68\" (UniqueName: \"kubernetes.io/projected/d63f3e93-f45f-4f7d-a29b-2eaae536fa5d-kube-api-access-xsq68\") pod \"d63f3e93-f45f-4f7d-a29b-2eaae536fa5d\" (UID: \"d63f3e93-f45f-4f7d-a29b-2eaae536fa5d\") " Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.169014 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d63f3e93-f45f-4f7d-a29b-2eaae536fa5d-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "d63f3e93-f45f-4f7d-a29b-2eaae536fa5d" (UID: "d63f3e93-f45f-4f7d-a29b-2eaae536fa5d"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.169048 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d63f3e93-f45f-4f7d-a29b-2eaae536fa5d-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "d63f3e93-f45f-4f7d-a29b-2eaae536fa5d" (UID: "d63f3e93-f45f-4f7d-a29b-2eaae536fa5d"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.169120 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d63f3e93-f45f-4f7d-a29b-2eaae536fa5d-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "d63f3e93-f45f-4f7d-a29b-2eaae536fa5d" (UID: "d63f3e93-f45f-4f7d-a29b-2eaae536fa5d"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.169690 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d63f3e93-f45f-4f7d-a29b-2eaae536fa5d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d63f3e93-f45f-4f7d-a29b-2eaae536fa5d" (UID: "d63f3e93-f45f-4f7d-a29b-2eaae536fa5d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.173280 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d63f3e93-f45f-4f7d-a29b-2eaae536fa5d-kube-api-access-xsq68" (OuterVolumeSpecName: "kube-api-access-xsq68") pod "d63f3e93-f45f-4f7d-a29b-2eaae536fa5d" (UID: "d63f3e93-f45f-4f7d-a29b-2eaae536fa5d"). InnerVolumeSpecName "kube-api-access-xsq68". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.177663 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "mysql-db") pod "d63f3e93-f45f-4f7d-a29b-2eaae536fa5d" (UID: "d63f3e93-f45f-4f7d-a29b-2eaae536fa5d"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.267911 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="316f3a85-5cf5-44c5-b8f5-fbc6d7cbff3d" path="/var/lib/kubelet/pods/316f3a85-5cf5-44c5-b8f5-fbc6d7cbff3d/volumes" Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.268395 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="353d8d01-e2c8-40ec-80e4-034ab7786569" path="/var/lib/kubelet/pods/353d8d01-e2c8-40ec-80e4-034ab7786569/volumes" Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.268844 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a972e8cd-cae8-4894-b2d3-7bfff75acc1c" path="/var/lib/kubelet/pods/a972e8cd-cae8-4894-b2d3-7bfff75acc1c/volumes" Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.269273 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8f2d3a6-ad9f-4e1c-8f9a-41f8c7efcdac" path="/var/lib/kubelet/pods/e8f2d3a6-ad9f-4e1c-8f9a-41f8c7efcdac/volumes" Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.270694 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d63f3e93-f45f-4f7d-a29b-2eaae536fa5d-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.270723 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xsq68\" (UniqueName: \"kubernetes.io/projected/d63f3e93-f45f-4f7d-a29b-2eaae536fa5d-kube-api-access-xsq68\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.270758 4733 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.270772 4733 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d63f3e93-f45f-4f7d-a29b-2eaae536fa5d-config-data-generated\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.270784 4733 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d63f3e93-f45f-4f7d-a29b-2eaae536fa5d-kolla-config\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.270796 4733 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d63f3e93-f45f-4f7d-a29b-2eaae536fa5d-config-data-default\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.277698 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f45e90dd-ab9d-4c27-8449-98cf52f48c87" path="/var/lib/kubelet/pods/f45e90dd-ab9d-4c27-8449-98cf52f48c87/volumes" Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.282024 4733 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.362616 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/keystonec686-account-delete-zqss2" Feb 02 15:38:45 crc kubenswrapper[4733]: E0202 15:38:45.371857 4733 configmap.go:193] Couldn't get configMap manila-kuttl-tests/openstack-scripts: configmap "openstack-scripts" not found Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.371902 4733 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:45 crc kubenswrapper[4733]: E0202 15:38:45.371920 4733 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/79f9eeab-dccd-4404-8179-fc3f776d8b22-operator-scripts podName:79f9eeab-dccd-4404-8179-fc3f776d8b22 nodeName:}" failed. No retries permitted until 2026-02-02 15:38:49.371906187 +0000 UTC m=+1412.823367545 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/79f9eeab-dccd-4404-8179-fc3f776d8b22-operator-scripts") pod "keystonec686-account-delete-zqss2" (UID: "79f9eeab-dccd-4404-8179-fc3f776d8b22") : configmap "openstack-scripts" not found Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.472860 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2mrmp\" (UniqueName: \"kubernetes.io/projected/79f9eeab-dccd-4404-8179-fc3f776d8b22-kube-api-access-2mrmp\") pod \"79f9eeab-dccd-4404-8179-fc3f776d8b22\" (UID: \"79f9eeab-dccd-4404-8179-fc3f776d8b22\") " Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.472958 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/79f9eeab-dccd-4404-8179-fc3f776d8b22-operator-scripts\") pod \"79f9eeab-dccd-4404-8179-fc3f776d8b22\" (UID: \"79f9eeab-dccd-4404-8179-fc3f776d8b22\") " Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.474753 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79f9eeab-dccd-4404-8179-fc3f776d8b22-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "79f9eeab-dccd-4404-8179-fc3f776d8b22" (UID: "79f9eeab-dccd-4404-8179-fc3f776d8b22"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.476923 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79f9eeab-dccd-4404-8179-fc3f776d8b22-kube-api-access-2mrmp" (OuterVolumeSpecName: "kube-api-access-2mrmp") pod "79f9eeab-dccd-4404-8179-fc3f776d8b22" (UID: "79f9eeab-dccd-4404-8179-fc3f776d8b22"). InnerVolumeSpecName "kube-api-access-2mrmp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.574521 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/79f9eeab-dccd-4404-8179-fc3f776d8b22-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.574555 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2mrmp\" (UniqueName: \"kubernetes.io/projected/79f9eeab-dccd-4404-8179-fc3f776d8b22-kube-api-access-2mrmp\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.815651 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/keystonec686-account-delete-zqss2" event={"ID":"79f9eeab-dccd-4404-8179-fc3f776d8b22","Type":"ContainerDied","Data":"81c30aa929390953e710af0da8ec697184872b5c3b9ac749b42d4051e0b2ad0b"} Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.815708 4733 scope.go:117] "RemoveContainer" containerID="5a580dc0dc11b40f3390a7c1a378ba8805983a64d2d3cedbeaf23f4a821f2126" Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.815669 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/keystonec686-account-delete-zqss2" Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.817672 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/openstack-galera-0" Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.817841 4733 generic.go:334] "Generic (PLEG): container finished" podID="f177ebf9-728b-4ba2-b9b7-6c676f9d014a" containerID="69b15a1f6699ebd76df18a6d10489a9a8f470c2c56ccab52ad3c08a533bb0a92" exitCode=0 Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.817898 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/openstack-galera-0" event={"ID":"f177ebf9-728b-4ba2-b9b7-6c676f9d014a","Type":"ContainerDied","Data":"69b15a1f6699ebd76df18a6d10489a9a8f470c2c56ccab52ad3c08a533bb0a92"} Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.817920 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/openstack-galera-0" event={"ID":"f177ebf9-728b-4ba2-b9b7-6c676f9d014a","Type":"ContainerDied","Data":"7c698d2bb4ec40a7f26727371b9ce3b5fbfc121e40731977721f74c904dbadf0"} Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.820352 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/openstack-galera-1" Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.820351 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/openstack-galera-1" event={"ID":"d63f3e93-f45f-4f7d-a29b-2eaae536fa5d","Type":"ContainerDied","Data":"c9a196e8e4c248e6bcc8d38449b00c75bf0edae74116bba0e25b9abf856f4a92"} Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.822693 4733 generic.go:334] "Generic (PLEG): container finished" podID="484de6bc-40c4-4159-a4f5-9cf345357b40" containerID="554cf0d17dd671921a0f9b08eecb0780c2e26ef19934f485452972fbf2dbf1f9" exitCode=0 Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.822805 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9gtt4" event={"ID":"484de6bc-40c4-4159-a4f5-9cf345357b40","Type":"ContainerDied","Data":"554cf0d17dd671921a0f9b08eecb0780c2e26ef19934f485452972fbf2dbf1f9"} Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.835995 4733 scope.go:117] "RemoveContainer" containerID="69b15a1f6699ebd76df18a6d10489a9a8f470c2c56ccab52ad3c08a533bb0a92" Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.865005 4733 scope.go:117] "RemoveContainer" containerID="6d16370fe61c34fb815b15b673736c3407ad87871969c2aab8e91a4ac281d254" Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.871903 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/keystonec686-account-delete-zqss2"] Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.878747 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"f177ebf9-728b-4ba2-b9b7-6c676f9d014a\" (UID: \"f177ebf9-728b-4ba2-b9b7-6c676f9d014a\") " Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.878899 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f177ebf9-728b-4ba2-b9b7-6c676f9d014a-config-data-generated\") pod \"f177ebf9-728b-4ba2-b9b7-6c676f9d014a\" (UID: \"f177ebf9-728b-4ba2-b9b7-6c676f9d014a\") " Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.879244 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f177ebf9-728b-4ba2-b9b7-6c676f9d014a-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "f177ebf9-728b-4ba2-b9b7-6c676f9d014a" (UID: "f177ebf9-728b-4ba2-b9b7-6c676f9d014a"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.879291 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f177ebf9-728b-4ba2-b9b7-6c676f9d014a-kolla-config\") pod \"f177ebf9-728b-4ba2-b9b7-6c676f9d014a\" (UID: \"f177ebf9-728b-4ba2-b9b7-6c676f9d014a\") " Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.879695 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f177ebf9-728b-4ba2-b9b7-6c676f9d014a-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "f177ebf9-728b-4ba2-b9b7-6c676f9d014a" (UID: "f177ebf9-728b-4ba2-b9b7-6c676f9d014a"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.879757 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f4kxk\" (UniqueName: \"kubernetes.io/projected/f177ebf9-728b-4ba2-b9b7-6c676f9d014a-kube-api-access-f4kxk\") pod \"f177ebf9-728b-4ba2-b9b7-6c676f9d014a\" (UID: \"f177ebf9-728b-4ba2-b9b7-6c676f9d014a\") " Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.879798 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f177ebf9-728b-4ba2-b9b7-6c676f9d014a-operator-scripts\") pod \"f177ebf9-728b-4ba2-b9b7-6c676f9d014a\" (UID: \"f177ebf9-728b-4ba2-b9b7-6c676f9d014a\") " Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.879836 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f177ebf9-728b-4ba2-b9b7-6c676f9d014a-config-data-default\") pod \"f177ebf9-728b-4ba2-b9b7-6c676f9d014a\" (UID: \"f177ebf9-728b-4ba2-b9b7-6c676f9d014a\") " Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.880147 4733 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f177ebf9-728b-4ba2-b9b7-6c676f9d014a-config-data-generated\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.880186 4733 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f177ebf9-728b-4ba2-b9b7-6c676f9d014a-kolla-config\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.880605 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f177ebf9-728b-4ba2-b9b7-6c676f9d014a-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "f177ebf9-728b-4ba2-b9b7-6c676f9d014a" (UID: "f177ebf9-728b-4ba2-b9b7-6c676f9d014a"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.881212 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["manila-kuttl-tests/keystonec686-account-delete-zqss2"] Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.881398 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f177ebf9-728b-4ba2-b9b7-6c676f9d014a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f177ebf9-728b-4ba2-b9b7-6c676f9d014a" (UID: "f177ebf9-728b-4ba2-b9b7-6c676f9d014a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.886903 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f177ebf9-728b-4ba2-b9b7-6c676f9d014a-kube-api-access-f4kxk" (OuterVolumeSpecName: "kube-api-access-f4kxk") pod "f177ebf9-728b-4ba2-b9b7-6c676f9d014a" (UID: "f177ebf9-728b-4ba2-b9b7-6c676f9d014a"). InnerVolumeSpecName "kube-api-access-f4kxk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.889225 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/openstack-galera-1"] Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.893342 4733 scope.go:117] "RemoveContainer" containerID="69b15a1f6699ebd76df18a6d10489a9a8f470c2c56ccab52ad3c08a533bb0a92" Feb 02 15:38:45 crc kubenswrapper[4733]: E0202 15:38:45.895668 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69b15a1f6699ebd76df18a6d10489a9a8f470c2c56ccab52ad3c08a533bb0a92\": container with ID starting with 69b15a1f6699ebd76df18a6d10489a9a8f470c2c56ccab52ad3c08a533bb0a92 not found: ID does not exist" containerID="69b15a1f6699ebd76df18a6d10489a9a8f470c2c56ccab52ad3c08a533bb0a92" Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.895719 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69b15a1f6699ebd76df18a6d10489a9a8f470c2c56ccab52ad3c08a533bb0a92"} err="failed to get container status \"69b15a1f6699ebd76df18a6d10489a9a8f470c2c56ccab52ad3c08a533bb0a92\": rpc error: code = NotFound desc = could not find container \"69b15a1f6699ebd76df18a6d10489a9a8f470c2c56ccab52ad3c08a533bb0a92\": container with ID starting with 69b15a1f6699ebd76df18a6d10489a9a8f470c2c56ccab52ad3c08a533bb0a92 not found: ID does not exist" Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.895799 4733 scope.go:117] "RemoveContainer" containerID="6d16370fe61c34fb815b15b673736c3407ad87871969c2aab8e91a4ac281d254" Feb 02 15:38:45 crc kubenswrapper[4733]: E0202 15:38:45.896495 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d16370fe61c34fb815b15b673736c3407ad87871969c2aab8e91a4ac281d254\": container with ID starting with 6d16370fe61c34fb815b15b673736c3407ad87871969c2aab8e91a4ac281d254 not found: ID does not exist" containerID="6d16370fe61c34fb815b15b673736c3407ad87871969c2aab8e91a4ac281d254" Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.896520 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d16370fe61c34fb815b15b673736c3407ad87871969c2aab8e91a4ac281d254"} err="failed to get container status \"6d16370fe61c34fb815b15b673736c3407ad87871969c2aab8e91a4ac281d254\": rpc error: code = NotFound desc = could not find container \"6d16370fe61c34fb815b15b673736c3407ad87871969c2aab8e91a4ac281d254\": container with ID starting with 6d16370fe61c34fb815b15b673736c3407ad87871969c2aab8e91a4ac281d254 not found: ID does not exist" Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.896537 4733 scope.go:117] "RemoveContainer" containerID="77d577f4756bae43064e83c2d18077bad1b459a16e5122f3a5125f7ac44f7514" Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.897254 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["manila-kuttl-tests/openstack-galera-1"] Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.904467 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "mysql-db") pod "f177ebf9-728b-4ba2-b9b7-6c676f9d014a" (UID: "f177ebf9-728b-4ba2-b9b7-6c676f9d014a"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.921297 4733 scope.go:117] "RemoveContainer" containerID="cde1ae39c9ba2c79c8411fde1fac0b652cd522d574af6fd6222107e900c130c5" Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.981096 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f4kxk\" (UniqueName: \"kubernetes.io/projected/f177ebf9-728b-4ba2-b9b7-6c676f9d014a-kube-api-access-f4kxk\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.981135 4733 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f177ebf9-728b-4ba2-b9b7-6c676f9d014a-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.981143 4733 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f177ebf9-728b-4ba2-b9b7-6c676f9d014a-config-data-default\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.981190 4733 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Feb 02 15:38:45 crc kubenswrapper[4733]: I0202 15:38:45.991929 4733 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Feb 02 15:38:46 crc kubenswrapper[4733]: I0202 15:38:46.082480 4733 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:46 crc kubenswrapper[4733]: I0202 15:38:46.838612 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/openstack-galera-0" Feb 02 15:38:46 crc kubenswrapper[4733]: I0202 15:38:46.875357 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/openstack-galera-0"] Feb 02 15:38:46 crc kubenswrapper[4733]: I0202 15:38:46.877569 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["manila-kuttl-tests/openstack-galera-0"] Feb 02 15:38:47 crc kubenswrapper[4733]: I0202 15:38:47.264463 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79f9eeab-dccd-4404-8179-fc3f776d8b22" path="/var/lib/kubelet/pods/79f9eeab-dccd-4404-8179-fc3f776d8b22/volumes" Feb 02 15:38:47 crc kubenswrapper[4733]: I0202 15:38:47.265546 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d63f3e93-f45f-4f7d-a29b-2eaae536fa5d" path="/var/lib/kubelet/pods/d63f3e93-f45f-4f7d-a29b-2eaae536fa5d/volumes" Feb 02 15:38:47 crc kubenswrapper[4733]: I0202 15:38:47.266443 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f177ebf9-728b-4ba2-b9b7-6c676f9d014a" path="/var/lib/kubelet/pods/f177ebf9-728b-4ba2-b9b7-6c676f9d014a/volumes" Feb 02 15:38:47 crc kubenswrapper[4733]: I0202 15:38:47.852094 4733 generic.go:334] "Generic (PLEG): container finished" podID="484de6bc-40c4-4159-a4f5-9cf345357b40" containerID="d45109ae4d569ea57dda3a5dec818ad03098c8f2bd31af4872e336b9347718d5" exitCode=0 Feb 02 15:38:47 crc kubenswrapper[4733]: I0202 15:38:47.852141 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9gtt4" event={"ID":"484de6bc-40c4-4159-a4f5-9cf345357b40","Type":"ContainerDied","Data":"d45109ae4d569ea57dda3a5dec818ad03098c8f2bd31af4872e336b9347718d5"} Feb 02 15:38:48 crc kubenswrapper[4733]: I0202 15:38:48.865686 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9gtt4" event={"ID":"484de6bc-40c4-4159-a4f5-9cf345357b40","Type":"ContainerStarted","Data":"07704e47e4b17166fde169dc220f41a289913f06c3cf380b8a92d9187408ef68"} Feb 02 15:38:48 crc kubenswrapper[4733]: I0202 15:38:48.893264 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9gtt4" podStartSLOduration=3.435646449 podStartE2EDuration="5.8932339s" podCreationTimestamp="2026-02-02 15:38:43 +0000 UTC" firstStartedPulling="2026-02-02 15:38:45.823861002 +0000 UTC m=+1409.275322360" lastFinishedPulling="2026-02-02 15:38:48.281448413 +0000 UTC m=+1411.732909811" observedRunningTime="2026-02-02 15:38:48.886404563 +0000 UTC m=+1412.337865921" watchObservedRunningTime="2026-02-02 15:38:48.8932339 +0000 UTC m=+1412.344695298" Feb 02 15:38:54 crc kubenswrapper[4733]: I0202 15:38:54.123534 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9gtt4" Feb 02 15:38:54 crc kubenswrapper[4733]: I0202 15:38:54.124313 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9gtt4" Feb 02 15:38:54 crc kubenswrapper[4733]: I0202 15:38:54.192533 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9gtt4" Feb 02 15:38:54 crc kubenswrapper[4733]: I0202 15:38:54.981930 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9gtt4" Feb 02 15:38:57 crc kubenswrapper[4733]: I0202 15:38:57.341217 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9gtt4"] Feb 02 15:38:57 crc kubenswrapper[4733]: I0202 15:38:57.341822 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9gtt4" podUID="484de6bc-40c4-4159-a4f5-9cf345357b40" containerName="registry-server" containerID="cri-o://07704e47e4b17166fde169dc220f41a289913f06c3cf380b8a92d9187408ef68" gracePeriod=2 Feb 02 15:38:57 crc kubenswrapper[4733]: I0202 15:38:57.837528 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9gtt4" Feb 02 15:38:57 crc kubenswrapper[4733]: I0202 15:38:57.950741 4733 generic.go:334] "Generic (PLEG): container finished" podID="484de6bc-40c4-4159-a4f5-9cf345357b40" containerID="07704e47e4b17166fde169dc220f41a289913f06c3cf380b8a92d9187408ef68" exitCode=0 Feb 02 15:38:57 crc kubenswrapper[4733]: I0202 15:38:57.950803 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9gtt4" event={"ID":"484de6bc-40c4-4159-a4f5-9cf345357b40","Type":"ContainerDied","Data":"07704e47e4b17166fde169dc220f41a289913f06c3cf380b8a92d9187408ef68"} Feb 02 15:38:57 crc kubenswrapper[4733]: I0202 15:38:57.950875 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9gtt4" Feb 02 15:38:57 crc kubenswrapper[4733]: I0202 15:38:57.951367 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9gtt4" event={"ID":"484de6bc-40c4-4159-a4f5-9cf345357b40","Type":"ContainerDied","Data":"18a6cbd09141c7a50c8db5f6e1be47576829bf56bccc2297ef239428b0767d08"} Feb 02 15:38:57 crc kubenswrapper[4733]: I0202 15:38:57.951701 4733 scope.go:117] "RemoveContainer" containerID="07704e47e4b17166fde169dc220f41a289913f06c3cf380b8a92d9187408ef68" Feb 02 15:38:57 crc kubenswrapper[4733]: I0202 15:38:57.969814 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/484de6bc-40c4-4159-a4f5-9cf345357b40-catalog-content\") pod \"484de6bc-40c4-4159-a4f5-9cf345357b40\" (UID: \"484de6bc-40c4-4159-a4f5-9cf345357b40\") " Feb 02 15:38:57 crc kubenswrapper[4733]: I0202 15:38:57.969945 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/484de6bc-40c4-4159-a4f5-9cf345357b40-utilities\") pod \"484de6bc-40c4-4159-a4f5-9cf345357b40\" (UID: \"484de6bc-40c4-4159-a4f5-9cf345357b40\") " Feb 02 15:38:57 crc kubenswrapper[4733]: I0202 15:38:57.970021 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qwwkk\" (UniqueName: \"kubernetes.io/projected/484de6bc-40c4-4159-a4f5-9cf345357b40-kube-api-access-qwwkk\") pod \"484de6bc-40c4-4159-a4f5-9cf345357b40\" (UID: \"484de6bc-40c4-4159-a4f5-9cf345357b40\") " Feb 02 15:38:57 crc kubenswrapper[4733]: I0202 15:38:57.972761 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/484de6bc-40c4-4159-a4f5-9cf345357b40-utilities" (OuterVolumeSpecName: "utilities") pod "484de6bc-40c4-4159-a4f5-9cf345357b40" (UID: "484de6bc-40c4-4159-a4f5-9cf345357b40"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:38:57 crc kubenswrapper[4733]: I0202 15:38:57.976415 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/484de6bc-40c4-4159-a4f5-9cf345357b40-kube-api-access-qwwkk" (OuterVolumeSpecName: "kube-api-access-qwwkk") pod "484de6bc-40c4-4159-a4f5-9cf345357b40" (UID: "484de6bc-40c4-4159-a4f5-9cf345357b40"). InnerVolumeSpecName "kube-api-access-qwwkk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:38:57 crc kubenswrapper[4733]: I0202 15:38:57.979444 4733 scope.go:117] "RemoveContainer" containerID="d45109ae4d569ea57dda3a5dec818ad03098c8f2bd31af4872e336b9347718d5" Feb 02 15:38:58 crc kubenswrapper[4733]: I0202 15:38:58.036047 4733 scope.go:117] "RemoveContainer" containerID="554cf0d17dd671921a0f9b08eecb0780c2e26ef19934f485452972fbf2dbf1f9" Feb 02 15:38:58 crc kubenswrapper[4733]: I0202 15:38:58.046350 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/484de6bc-40c4-4159-a4f5-9cf345357b40-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "484de6bc-40c4-4159-a4f5-9cf345357b40" (UID: "484de6bc-40c4-4159-a4f5-9cf345357b40"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:38:58 crc kubenswrapper[4733]: I0202 15:38:58.061359 4733 scope.go:117] "RemoveContainer" containerID="07704e47e4b17166fde169dc220f41a289913f06c3cf380b8a92d9187408ef68" Feb 02 15:38:58 crc kubenswrapper[4733]: E0202 15:38:58.061808 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07704e47e4b17166fde169dc220f41a289913f06c3cf380b8a92d9187408ef68\": container with ID starting with 07704e47e4b17166fde169dc220f41a289913f06c3cf380b8a92d9187408ef68 not found: ID does not exist" containerID="07704e47e4b17166fde169dc220f41a289913f06c3cf380b8a92d9187408ef68" Feb 02 15:38:58 crc kubenswrapper[4733]: I0202 15:38:58.061845 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07704e47e4b17166fde169dc220f41a289913f06c3cf380b8a92d9187408ef68"} err="failed to get container status \"07704e47e4b17166fde169dc220f41a289913f06c3cf380b8a92d9187408ef68\": rpc error: code = NotFound desc = could not find container \"07704e47e4b17166fde169dc220f41a289913f06c3cf380b8a92d9187408ef68\": container with ID starting with 07704e47e4b17166fde169dc220f41a289913f06c3cf380b8a92d9187408ef68 not found: ID does not exist" Feb 02 15:38:58 crc kubenswrapper[4733]: I0202 15:38:58.061884 4733 scope.go:117] "RemoveContainer" containerID="d45109ae4d569ea57dda3a5dec818ad03098c8f2bd31af4872e336b9347718d5" Feb 02 15:38:58 crc kubenswrapper[4733]: E0202 15:38:58.062367 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d45109ae4d569ea57dda3a5dec818ad03098c8f2bd31af4872e336b9347718d5\": container with ID starting with d45109ae4d569ea57dda3a5dec818ad03098c8f2bd31af4872e336b9347718d5 not found: ID does not exist" containerID="d45109ae4d569ea57dda3a5dec818ad03098c8f2bd31af4872e336b9347718d5" Feb 02 15:38:58 crc kubenswrapper[4733]: I0202 15:38:58.062418 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d45109ae4d569ea57dda3a5dec818ad03098c8f2bd31af4872e336b9347718d5"} err="failed to get container status \"d45109ae4d569ea57dda3a5dec818ad03098c8f2bd31af4872e336b9347718d5\": rpc error: code = NotFound desc = could not find container \"d45109ae4d569ea57dda3a5dec818ad03098c8f2bd31af4872e336b9347718d5\": container with ID starting with d45109ae4d569ea57dda3a5dec818ad03098c8f2bd31af4872e336b9347718d5 not found: ID does not exist" Feb 02 15:38:58 crc kubenswrapper[4733]: I0202 15:38:58.062455 4733 scope.go:117] "RemoveContainer" containerID="554cf0d17dd671921a0f9b08eecb0780c2e26ef19934f485452972fbf2dbf1f9" Feb 02 15:38:58 crc kubenswrapper[4733]: E0202 15:38:58.062793 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"554cf0d17dd671921a0f9b08eecb0780c2e26ef19934f485452972fbf2dbf1f9\": container with ID starting with 554cf0d17dd671921a0f9b08eecb0780c2e26ef19934f485452972fbf2dbf1f9 not found: ID does not exist" containerID="554cf0d17dd671921a0f9b08eecb0780c2e26ef19934f485452972fbf2dbf1f9" Feb 02 15:38:58 crc kubenswrapper[4733]: I0202 15:38:58.062818 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"554cf0d17dd671921a0f9b08eecb0780c2e26ef19934f485452972fbf2dbf1f9"} err="failed to get container status \"554cf0d17dd671921a0f9b08eecb0780c2e26ef19934f485452972fbf2dbf1f9\": rpc error: code = NotFound desc = could not find container \"554cf0d17dd671921a0f9b08eecb0780c2e26ef19934f485452972fbf2dbf1f9\": container with ID starting with 554cf0d17dd671921a0f9b08eecb0780c2e26ef19934f485452972fbf2dbf1f9 not found: ID does not exist" Feb 02 15:38:58 crc kubenswrapper[4733]: I0202 15:38:58.071702 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qwwkk\" (UniqueName: \"kubernetes.io/projected/484de6bc-40c4-4159-a4f5-9cf345357b40-kube-api-access-qwwkk\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:58 crc kubenswrapper[4733]: I0202 15:38:58.071740 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/484de6bc-40c4-4159-a4f5-9cf345357b40-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:58 crc kubenswrapper[4733]: I0202 15:38:58.071752 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/484de6bc-40c4-4159-a4f5-9cf345357b40-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 15:38:58 crc kubenswrapper[4733]: I0202 15:38:58.300155 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9gtt4"] Feb 02 15:38:58 crc kubenswrapper[4733]: I0202 15:38:58.306857 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9gtt4"] Feb 02 15:38:59 crc kubenswrapper[4733]: I0202 15:38:59.273314 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="484de6bc-40c4-4159-a4f5-9cf345357b40" path="/var/lib/kubelet/pods/484de6bc-40c4-4159-a4f5-9cf345357b40/volumes" Feb 02 15:39:04 crc kubenswrapper[4733]: I0202 15:39:04.965323 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-service-cleanup-n5b5h655-8d2b7" Feb 02 15:39:04 crc kubenswrapper[4733]: I0202 15:39:04.987752 4733 patch_prober.go:28] interesting pod/machine-config-daemon-zgsv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 15:39:04 crc kubenswrapper[4733]: I0202 15:39:04.987813 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 15:39:05 crc kubenswrapper[4733]: I0202 15:39:05.023986 4733 generic.go:334] "Generic (PLEG): container finished" podID="9dffed5a-a57d-4710-909f-799b47ff68ef" containerID="ad443ac0179f4c8b1942430fb1ee0111ff937d00f7002251606b054c74e1138e" exitCode=137 Feb 02 15:39:05 crc kubenswrapper[4733]: I0202 15:39:05.024026 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-service-cleanup-n5b5h655-8d2b7" event={"ID":"9dffed5a-a57d-4710-909f-799b47ff68ef","Type":"ContainerDied","Data":"ad443ac0179f4c8b1942430fb1ee0111ff937d00f7002251606b054c74e1138e"} Feb 02 15:39:05 crc kubenswrapper[4733]: I0202 15:39:05.024076 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/manila-service-cleanup-n5b5h655-8d2b7" event={"ID":"9dffed5a-a57d-4710-909f-799b47ff68ef","Type":"ContainerDied","Data":"75d93405a5f061a3a7692dc88c34cbd68d03aafedc47ed7489f01c130550f0ec"} Feb 02 15:39:05 crc kubenswrapper[4733]: I0202 15:39:05.024094 4733 scope.go:117] "RemoveContainer" containerID="ad443ac0179f4c8b1942430fb1ee0111ff937d00f7002251606b054c74e1138e" Feb 02 15:39:05 crc kubenswrapper[4733]: I0202 15:39:05.024115 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/manila-service-cleanup-n5b5h655-8d2b7" Feb 02 15:39:05 crc kubenswrapper[4733]: I0202 15:39:05.053995 4733 scope.go:117] "RemoveContainer" containerID="ad443ac0179f4c8b1942430fb1ee0111ff937d00f7002251606b054c74e1138e" Feb 02 15:39:05 crc kubenswrapper[4733]: E0202 15:39:05.054599 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad443ac0179f4c8b1942430fb1ee0111ff937d00f7002251606b054c74e1138e\": container with ID starting with ad443ac0179f4c8b1942430fb1ee0111ff937d00f7002251606b054c74e1138e not found: ID does not exist" containerID="ad443ac0179f4c8b1942430fb1ee0111ff937d00f7002251606b054c74e1138e" Feb 02 15:39:05 crc kubenswrapper[4733]: I0202 15:39:05.054807 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad443ac0179f4c8b1942430fb1ee0111ff937d00f7002251606b054c74e1138e"} err="failed to get container status \"ad443ac0179f4c8b1942430fb1ee0111ff937d00f7002251606b054c74e1138e\": rpc error: code = NotFound desc = could not find container \"ad443ac0179f4c8b1942430fb1ee0111ff937d00f7002251606b054c74e1138e\": container with ID starting with ad443ac0179f4c8b1942430fb1ee0111ff937d00f7002251606b054c74e1138e not found: ID does not exist" Feb 02 15:39:05 crc kubenswrapper[4733]: I0202 15:39:05.080629 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8lms2\" (UniqueName: \"kubernetes.io/projected/9dffed5a-a57d-4710-909f-799b47ff68ef-kube-api-access-8lms2\") pod \"9dffed5a-a57d-4710-909f-799b47ff68ef\" (UID: \"9dffed5a-a57d-4710-909f-799b47ff68ef\") " Feb 02 15:39:05 crc kubenswrapper[4733]: I0202 15:39:05.080794 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9dffed5a-a57d-4710-909f-799b47ff68ef-config-data\") pod \"9dffed5a-a57d-4710-909f-799b47ff68ef\" (UID: \"9dffed5a-a57d-4710-909f-799b47ff68ef\") " Feb 02 15:39:05 crc kubenswrapper[4733]: I0202 15:39:05.080835 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/9dffed5a-a57d-4710-909f-799b47ff68ef-job-config-data\") pod \"9dffed5a-a57d-4710-909f-799b47ff68ef\" (UID: \"9dffed5a-a57d-4710-909f-799b47ff68ef\") " Feb 02 15:39:05 crc kubenswrapper[4733]: I0202 15:39:05.092353 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9dffed5a-a57d-4710-909f-799b47ff68ef-job-config-data" (OuterVolumeSpecName: "job-config-data") pod "9dffed5a-a57d-4710-909f-799b47ff68ef" (UID: "9dffed5a-a57d-4710-909f-799b47ff68ef"). InnerVolumeSpecName "job-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:39:05 crc kubenswrapper[4733]: I0202 15:39:05.092421 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9dffed5a-a57d-4710-909f-799b47ff68ef-kube-api-access-8lms2" (OuterVolumeSpecName: "kube-api-access-8lms2") pod "9dffed5a-a57d-4710-909f-799b47ff68ef" (UID: "9dffed5a-a57d-4710-909f-799b47ff68ef"). InnerVolumeSpecName "kube-api-access-8lms2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:39:05 crc kubenswrapper[4733]: I0202 15:39:05.096699 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9dffed5a-a57d-4710-909f-799b47ff68ef-config-data" (OuterVolumeSpecName: "config-data") pod "9dffed5a-a57d-4710-909f-799b47ff68ef" (UID: "9dffed5a-a57d-4710-909f-799b47ff68ef"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:39:05 crc kubenswrapper[4733]: I0202 15:39:05.183107 4733 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9dffed5a-a57d-4710-909f-799b47ff68ef-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 15:39:05 crc kubenswrapper[4733]: I0202 15:39:05.183195 4733 reconciler_common.go:293] "Volume detached for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/9dffed5a-a57d-4710-909f-799b47ff68ef-job-config-data\") on node \"crc\" DevicePath \"\"" Feb 02 15:39:05 crc kubenswrapper[4733]: I0202 15:39:05.183229 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8lms2\" (UniqueName: \"kubernetes.io/projected/9dffed5a-a57d-4710-909f-799b47ff68ef-kube-api-access-8lms2\") on node \"crc\" DevicePath \"\"" Feb 02 15:39:05 crc kubenswrapper[4733]: I0202 15:39:05.351845 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/manila-service-cleanup-n5b5h655-8d2b7"] Feb 02 15:39:05 crc kubenswrapper[4733]: I0202 15:39:05.360032 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["manila-kuttl-tests/manila-service-cleanup-n5b5h655-8d2b7"] Feb 02 15:39:07 crc kubenswrapper[4733]: I0202 15:39:07.270539 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9dffed5a-a57d-4710-909f-799b47ff68ef" path="/var/lib/kubelet/pods/9dffed5a-a57d-4710-909f-799b47ff68ef/volumes" Feb 02 15:39:13 crc kubenswrapper[4733]: I0202 15:39:13.107768 4733 generic.go:334] "Generic (PLEG): container finished" podID="c71eca7f-f211-40bc-9116-f3c246c46e54" containerID="00f13c6c3710736a5fd10709601532b0ab4f0709341f931f4399424374017bce" exitCode=137 Feb 02 15:39:13 crc kubenswrapper[4733]: I0202 15:39:13.107846 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/ceph" event={"ID":"c71eca7f-f211-40bc-9116-f3c246c46e54","Type":"ContainerDied","Data":"00f13c6c3710736a5fd10709601532b0ab4f0709341f931f4399424374017bce"} Feb 02 15:39:13 crc kubenswrapper[4733]: I0202 15:39:13.400317 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/ceph" Feb 02 15:39:13 crc kubenswrapper[4733]: I0202 15:39:13.517536 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pc87n\" (UniqueName: \"kubernetes.io/projected/c71eca7f-f211-40bc-9116-f3c246c46e54-kube-api-access-pc87n\") pod \"c71eca7f-f211-40bc-9116-f3c246c46e54\" (UID: \"c71eca7f-f211-40bc-9116-f3c246c46e54\") " Feb 02 15:39:13 crc kubenswrapper[4733]: I0202 15:39:13.517593 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log\" (UniqueName: \"kubernetes.io/empty-dir/c71eca7f-f211-40bc-9116-f3c246c46e54-log\") pod \"c71eca7f-f211-40bc-9116-f3c246c46e54\" (UID: \"c71eca7f-f211-40bc-9116-f3c246c46e54\") " Feb 02 15:39:13 crc kubenswrapper[4733]: I0202 15:39:13.517656 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/empty-dir/c71eca7f-f211-40bc-9116-f3c246c46e54-run\") pod \"c71eca7f-f211-40bc-9116-f3c246c46e54\" (UID: \"c71eca7f-f211-40bc-9116-f3c246c46e54\") " Feb 02 15:39:13 crc kubenswrapper[4733]: I0202 15:39:13.517682 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"data\" (UniqueName: \"kubernetes.io/empty-dir/c71eca7f-f211-40bc-9116-f3c246c46e54-data\") pod \"c71eca7f-f211-40bc-9116-f3c246c46e54\" (UID: \"c71eca7f-f211-40bc-9116-f3c246c46e54\") " Feb 02 15:39:13 crc kubenswrapper[4733]: I0202 15:39:13.518733 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c71eca7f-f211-40bc-9116-f3c246c46e54-run" (OuterVolumeSpecName: "run") pod "c71eca7f-f211-40bc-9116-f3c246c46e54" (UID: "c71eca7f-f211-40bc-9116-f3c246c46e54"). InnerVolumeSpecName "run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:39:13 crc kubenswrapper[4733]: I0202 15:39:13.518924 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c71eca7f-f211-40bc-9116-f3c246c46e54-log" (OuterVolumeSpecName: "log") pod "c71eca7f-f211-40bc-9116-f3c246c46e54" (UID: "c71eca7f-f211-40bc-9116-f3c246c46e54"). InnerVolumeSpecName "log". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:39:13 crc kubenswrapper[4733]: I0202 15:39:13.523908 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c71eca7f-f211-40bc-9116-f3c246c46e54-data" (OuterVolumeSpecName: "data") pod "c71eca7f-f211-40bc-9116-f3c246c46e54" (UID: "c71eca7f-f211-40bc-9116-f3c246c46e54"). InnerVolumeSpecName "data". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:39:13 crc kubenswrapper[4733]: I0202 15:39:13.526626 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c71eca7f-f211-40bc-9116-f3c246c46e54-kube-api-access-pc87n" (OuterVolumeSpecName: "kube-api-access-pc87n") pod "c71eca7f-f211-40bc-9116-f3c246c46e54" (UID: "c71eca7f-f211-40bc-9116-f3c246c46e54"). InnerVolumeSpecName "kube-api-access-pc87n". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:39:13 crc kubenswrapper[4733]: I0202 15:39:13.618867 4733 reconciler_common.go:293] "Volume detached for volume \"log\" (UniqueName: \"kubernetes.io/empty-dir/c71eca7f-f211-40bc-9116-f3c246c46e54-log\") on node \"crc\" DevicePath \"\"" Feb 02 15:39:13 crc kubenswrapper[4733]: I0202 15:39:13.618902 4733 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/empty-dir/c71eca7f-f211-40bc-9116-f3c246c46e54-run\") on node \"crc\" DevicePath \"\"" Feb 02 15:39:13 crc kubenswrapper[4733]: I0202 15:39:13.618917 4733 reconciler_common.go:293] "Volume detached for volume \"data\" (UniqueName: \"kubernetes.io/empty-dir/c71eca7f-f211-40bc-9116-f3c246c46e54-data\") on node \"crc\" DevicePath \"\"" Feb 02 15:39:13 crc kubenswrapper[4733]: I0202 15:39:13.618930 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pc87n\" (UniqueName: \"kubernetes.io/projected/c71eca7f-f211-40bc-9116-f3c246c46e54-kube-api-access-pc87n\") on node \"crc\" DevicePath \"\"" Feb 02 15:39:14 crc kubenswrapper[4733]: I0202 15:39:14.122521 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="manila-kuttl-tests/ceph" event={"ID":"c71eca7f-f211-40bc-9116-f3c246c46e54","Type":"ContainerDied","Data":"ac9468187bfa0e717e4d13ba5017a4db3306b2c3ae4955acf6cc91b309ae3246"} Feb 02 15:39:14 crc kubenswrapper[4733]: I0202 15:39:14.122583 4733 scope.go:117] "RemoveContainer" containerID="00f13c6c3710736a5fd10709601532b0ab4f0709341f931f4399424374017bce" Feb 02 15:39:14 crc kubenswrapper[4733]: I0202 15:39:14.122632 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="manila-kuttl-tests/ceph" Feb 02 15:39:14 crc kubenswrapper[4733]: I0202 15:39:14.223236 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["manila-kuttl-tests/ceph"] Feb 02 15:39:14 crc kubenswrapper[4733]: I0202 15:39:14.227434 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["manila-kuttl-tests/ceph"] Feb 02 15:39:15 crc kubenswrapper[4733]: I0202 15:39:15.025248 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-86bf5ddb6d-lx2j7"] Feb 02 15:39:15 crc kubenswrapper[4733]: I0202 15:39:15.025445 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/keystone-operator-controller-manager-86bf5ddb6d-lx2j7" podUID="1aaf01fd-0114-41a6-af39-bd1a10deabe9" containerName="manager" containerID="cri-o://a81e99b0593d8b4fc3e8efc759cf74ef72b91cef1303d352fdc881f4c7aa0104" gracePeriod=10 Feb 02 15:39:15 crc kubenswrapper[4733]: I0202 15:39:15.288805 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c71eca7f-f211-40bc-9116-f3c246c46e54" path="/var/lib/kubelet/pods/c71eca7f-f211-40bc-9116-f3c246c46e54/volumes" Feb 02 15:39:15 crc kubenswrapper[4733]: I0202 15:39:15.360189 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-index-l4jcf"] Feb 02 15:39:15 crc kubenswrapper[4733]: I0202 15:39:15.360372 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/keystone-operator-index-l4jcf" podUID="d0147e51-446d-4a71-aaaa-e852258495d4" containerName="registry-server" containerID="cri-o://e8fd5017367c208f4e3389d15fb93e2782229ba4ab3d82f8fefe713594ebcad1" gracePeriod=30 Feb 02 15:39:15 crc kubenswrapper[4733]: I0202 15:39:15.397629 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqw8xf"] Feb 02 15:39:15 crc kubenswrapper[4733]: I0202 15:39:15.402838 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/b43f19b8e3bb8997a527070b172ae030accff9cd1a2f2b076f58d9c4efqw8xf"] Feb 02 15:39:15 crc kubenswrapper[4733]: I0202 15:39:15.492877 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-86bf5ddb6d-lx2j7" Feb 02 15:39:15 crc kubenswrapper[4733]: I0202 15:39:15.653558 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4qzm\" (UniqueName: \"kubernetes.io/projected/1aaf01fd-0114-41a6-af39-bd1a10deabe9-kube-api-access-d4qzm\") pod \"1aaf01fd-0114-41a6-af39-bd1a10deabe9\" (UID: \"1aaf01fd-0114-41a6-af39-bd1a10deabe9\") " Feb 02 15:39:15 crc kubenswrapper[4733]: I0202 15:39:15.653699 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1aaf01fd-0114-41a6-af39-bd1a10deabe9-apiservice-cert\") pod \"1aaf01fd-0114-41a6-af39-bd1a10deabe9\" (UID: \"1aaf01fd-0114-41a6-af39-bd1a10deabe9\") " Feb 02 15:39:15 crc kubenswrapper[4733]: I0202 15:39:15.653814 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1aaf01fd-0114-41a6-af39-bd1a10deabe9-webhook-cert\") pod \"1aaf01fd-0114-41a6-af39-bd1a10deabe9\" (UID: \"1aaf01fd-0114-41a6-af39-bd1a10deabe9\") " Feb 02 15:39:15 crc kubenswrapper[4733]: I0202 15:39:15.659508 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1aaf01fd-0114-41a6-af39-bd1a10deabe9-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "1aaf01fd-0114-41a6-af39-bd1a10deabe9" (UID: "1aaf01fd-0114-41a6-af39-bd1a10deabe9"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:39:15 crc kubenswrapper[4733]: I0202 15:39:15.661382 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1aaf01fd-0114-41a6-af39-bd1a10deabe9-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "1aaf01fd-0114-41a6-af39-bd1a10deabe9" (UID: "1aaf01fd-0114-41a6-af39-bd1a10deabe9"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:39:15 crc kubenswrapper[4733]: I0202 15:39:15.667296 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1aaf01fd-0114-41a6-af39-bd1a10deabe9-kube-api-access-d4qzm" (OuterVolumeSpecName: "kube-api-access-d4qzm") pod "1aaf01fd-0114-41a6-af39-bd1a10deabe9" (UID: "1aaf01fd-0114-41a6-af39-bd1a10deabe9"). InnerVolumeSpecName "kube-api-access-d4qzm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:39:15 crc kubenswrapper[4733]: I0202 15:39:15.718340 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-l4jcf" Feb 02 15:39:15 crc kubenswrapper[4733]: I0202 15:39:15.762534 4733 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1aaf01fd-0114-41a6-af39-bd1a10deabe9-webhook-cert\") on node \"crc\" DevicePath \"\"" Feb 02 15:39:15 crc kubenswrapper[4733]: I0202 15:39:15.762566 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4qzm\" (UniqueName: \"kubernetes.io/projected/1aaf01fd-0114-41a6-af39-bd1a10deabe9-kube-api-access-d4qzm\") on node \"crc\" DevicePath \"\"" Feb 02 15:39:15 crc kubenswrapper[4733]: I0202 15:39:15.762576 4733 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1aaf01fd-0114-41a6-af39-bd1a10deabe9-apiservice-cert\") on node \"crc\" DevicePath \"\"" Feb 02 15:39:15 crc kubenswrapper[4733]: I0202 15:39:15.864023 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dl76v\" (UniqueName: \"kubernetes.io/projected/d0147e51-446d-4a71-aaaa-e852258495d4-kube-api-access-dl76v\") pod \"d0147e51-446d-4a71-aaaa-e852258495d4\" (UID: \"d0147e51-446d-4a71-aaaa-e852258495d4\") " Feb 02 15:39:15 crc kubenswrapper[4733]: I0202 15:39:15.866553 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0147e51-446d-4a71-aaaa-e852258495d4-kube-api-access-dl76v" (OuterVolumeSpecName: "kube-api-access-dl76v") pod "d0147e51-446d-4a71-aaaa-e852258495d4" (UID: "d0147e51-446d-4a71-aaaa-e852258495d4"). InnerVolumeSpecName "kube-api-access-dl76v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:39:15 crc kubenswrapper[4733]: I0202 15:39:15.965967 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dl76v\" (UniqueName: \"kubernetes.io/projected/d0147e51-446d-4a71-aaaa-e852258495d4-kube-api-access-dl76v\") on node \"crc\" DevicePath \"\"" Feb 02 15:39:16 crc kubenswrapper[4733]: I0202 15:39:16.143583 4733 generic.go:334] "Generic (PLEG): container finished" podID="1aaf01fd-0114-41a6-af39-bd1a10deabe9" containerID="a81e99b0593d8b4fc3e8efc759cf74ef72b91cef1303d352fdc881f4c7aa0104" exitCode=0 Feb 02 15:39:16 crc kubenswrapper[4733]: I0202 15:39:16.143661 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-86bf5ddb6d-lx2j7" Feb 02 15:39:16 crc kubenswrapper[4733]: I0202 15:39:16.143668 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-86bf5ddb6d-lx2j7" event={"ID":"1aaf01fd-0114-41a6-af39-bd1a10deabe9","Type":"ContainerDied","Data":"a81e99b0593d8b4fc3e8efc759cf74ef72b91cef1303d352fdc881f4c7aa0104"} Feb 02 15:39:16 crc kubenswrapper[4733]: I0202 15:39:16.144227 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-86bf5ddb6d-lx2j7" event={"ID":"1aaf01fd-0114-41a6-af39-bd1a10deabe9","Type":"ContainerDied","Data":"994566ca492a57f666784fc02b02c3ef6f5012b842546007a5985ae6268d8324"} Feb 02 15:39:16 crc kubenswrapper[4733]: I0202 15:39:16.144264 4733 scope.go:117] "RemoveContainer" containerID="a81e99b0593d8b4fc3e8efc759cf74ef72b91cef1303d352fdc881f4c7aa0104" Feb 02 15:39:16 crc kubenswrapper[4733]: I0202 15:39:16.151811 4733 generic.go:334] "Generic (PLEG): container finished" podID="d0147e51-446d-4a71-aaaa-e852258495d4" containerID="e8fd5017367c208f4e3389d15fb93e2782229ba4ab3d82f8fefe713594ebcad1" exitCode=0 Feb 02 15:39:16 crc kubenswrapper[4733]: I0202 15:39:16.151868 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-l4jcf" event={"ID":"d0147e51-446d-4a71-aaaa-e852258495d4","Type":"ContainerDied","Data":"e8fd5017367c208f4e3389d15fb93e2782229ba4ab3d82f8fefe713594ebcad1"} Feb 02 15:39:16 crc kubenswrapper[4733]: I0202 15:39:16.151901 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-l4jcf" event={"ID":"d0147e51-446d-4a71-aaaa-e852258495d4","Type":"ContainerDied","Data":"5545c2b365737082b809c4c8d0bc0c975465df9799553b7895fda48062e0eecb"} Feb 02 15:39:16 crc kubenswrapper[4733]: I0202 15:39:16.151921 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-l4jcf" Feb 02 15:39:16 crc kubenswrapper[4733]: I0202 15:39:16.208900 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-86bf5ddb6d-lx2j7"] Feb 02 15:39:16 crc kubenswrapper[4733]: I0202 15:39:16.218197 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-86bf5ddb6d-lx2j7"] Feb 02 15:39:16 crc kubenswrapper[4733]: I0202 15:39:16.221935 4733 scope.go:117] "RemoveContainer" containerID="a81e99b0593d8b4fc3e8efc759cf74ef72b91cef1303d352fdc881f4c7aa0104" Feb 02 15:39:16 crc kubenswrapper[4733]: E0202 15:39:16.223279 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a81e99b0593d8b4fc3e8efc759cf74ef72b91cef1303d352fdc881f4c7aa0104\": container with ID starting with a81e99b0593d8b4fc3e8efc759cf74ef72b91cef1303d352fdc881f4c7aa0104 not found: ID does not exist" containerID="a81e99b0593d8b4fc3e8efc759cf74ef72b91cef1303d352fdc881f4c7aa0104" Feb 02 15:39:16 crc kubenswrapper[4733]: I0202 15:39:16.223331 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a81e99b0593d8b4fc3e8efc759cf74ef72b91cef1303d352fdc881f4c7aa0104"} err="failed to get container status \"a81e99b0593d8b4fc3e8efc759cf74ef72b91cef1303d352fdc881f4c7aa0104\": rpc error: code = NotFound desc = could not find container \"a81e99b0593d8b4fc3e8efc759cf74ef72b91cef1303d352fdc881f4c7aa0104\": container with ID starting with a81e99b0593d8b4fc3e8efc759cf74ef72b91cef1303d352fdc881f4c7aa0104 not found: ID does not exist" Feb 02 15:39:16 crc kubenswrapper[4733]: I0202 15:39:16.223371 4733 scope.go:117] "RemoveContainer" containerID="e8fd5017367c208f4e3389d15fb93e2782229ba4ab3d82f8fefe713594ebcad1" Feb 02 15:39:16 crc kubenswrapper[4733]: I0202 15:39:16.228657 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-index-l4jcf"] Feb 02 15:39:16 crc kubenswrapper[4733]: I0202 15:39:16.241244 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/keystone-operator-index-l4jcf"] Feb 02 15:39:16 crc kubenswrapper[4733]: I0202 15:39:16.251988 4733 scope.go:117] "RemoveContainer" containerID="e8fd5017367c208f4e3389d15fb93e2782229ba4ab3d82f8fefe713594ebcad1" Feb 02 15:39:16 crc kubenswrapper[4733]: E0202 15:39:16.252855 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8fd5017367c208f4e3389d15fb93e2782229ba4ab3d82f8fefe713594ebcad1\": container with ID starting with e8fd5017367c208f4e3389d15fb93e2782229ba4ab3d82f8fefe713594ebcad1 not found: ID does not exist" containerID="e8fd5017367c208f4e3389d15fb93e2782229ba4ab3d82f8fefe713594ebcad1" Feb 02 15:39:16 crc kubenswrapper[4733]: I0202 15:39:16.252906 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8fd5017367c208f4e3389d15fb93e2782229ba4ab3d82f8fefe713594ebcad1"} err="failed to get container status \"e8fd5017367c208f4e3389d15fb93e2782229ba4ab3d82f8fefe713594ebcad1\": rpc error: code = NotFound desc = could not find container \"e8fd5017367c208f4e3389d15fb93e2782229ba4ab3d82f8fefe713594ebcad1\": container with ID starting with e8fd5017367c208f4e3389d15fb93e2782229ba4ab3d82f8fefe713594ebcad1 not found: ID does not exist" Feb 02 15:39:17 crc kubenswrapper[4733]: I0202 15:39:17.266643 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1aaf01fd-0114-41a6-af39-bd1a10deabe9" path="/var/lib/kubelet/pods/1aaf01fd-0114-41a6-af39-bd1a10deabe9/volumes" Feb 02 15:39:17 crc kubenswrapper[4733]: I0202 15:39:17.267124 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fbae2a6-73a0-496e-900c-6d0d4c704994" path="/var/lib/kubelet/pods/2fbae2a6-73a0-496e-900c-6d0d4c704994/volumes" Feb 02 15:39:17 crc kubenswrapper[4733]: I0202 15:39:17.267662 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0147e51-446d-4a71-aaaa-e852258495d4" path="/var/lib/kubelet/pods/d0147e51-446d-4a71-aaaa-e852258495d4/volumes" Feb 02 15:39:18 crc kubenswrapper[4733]: I0202 15:39:18.311575 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-gzhj7"] Feb 02 15:39:18 crc kubenswrapper[4733]: I0202 15:39:18.312035 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-gzhj7" podUID="2e4d2825-196e-4bf2-8f47-270645f2267b" containerName="operator" containerID="cri-o://99ee0047c1f019d60cd3d61bf905abf882cb05edbb2b3e61679da900a5465bc4" gracePeriod=10 Feb 02 15:39:18 crc kubenswrapper[4733]: I0202 15:39:18.603639 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-nrbdq"] Feb 02 15:39:18 crc kubenswrapper[4733]: I0202 15:39:18.603824 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/rabbitmq-cluster-operator-index-nrbdq" podUID="855c71f2-a025-49cb-84c8-0206b2ec7e08" containerName="registry-server" containerID="cri-o://ba22d52e2c88845b48608d81df18d063e6040d2d90a6f634187534bf7ca77329" gracePeriod=30 Feb 02 15:39:18 crc kubenswrapper[4733]: I0202 15:39:18.636798 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590wfw27"] Feb 02 15:39:18 crc kubenswrapper[4733]: I0202 15:39:18.642585 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590wfw27"] Feb 02 15:39:18 crc kubenswrapper[4733]: I0202 15:39:18.747080 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-gzhj7" Feb 02 15:39:18 crc kubenswrapper[4733]: I0202 15:39:18.806548 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lrcrm\" (UniqueName: \"kubernetes.io/projected/2e4d2825-196e-4bf2-8f47-270645f2267b-kube-api-access-lrcrm\") pod \"2e4d2825-196e-4bf2-8f47-270645f2267b\" (UID: \"2e4d2825-196e-4bf2-8f47-270645f2267b\") " Feb 02 15:39:18 crc kubenswrapper[4733]: I0202 15:39:18.812357 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e4d2825-196e-4bf2-8f47-270645f2267b-kube-api-access-lrcrm" (OuterVolumeSpecName: "kube-api-access-lrcrm") pod "2e4d2825-196e-4bf2-8f47-270645f2267b" (UID: "2e4d2825-196e-4bf2-8f47-270645f2267b"). InnerVolumeSpecName "kube-api-access-lrcrm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:39:18 crc kubenswrapper[4733]: I0202 15:39:18.908634 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lrcrm\" (UniqueName: \"kubernetes.io/projected/2e4d2825-196e-4bf2-8f47-270645f2267b-kube-api-access-lrcrm\") on node \"crc\" DevicePath \"\"" Feb 02 15:39:19 crc kubenswrapper[4733]: I0202 15:39:19.057558 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-nrbdq" Feb 02 15:39:19 crc kubenswrapper[4733]: I0202 15:39:19.185400 4733 generic.go:334] "Generic (PLEG): container finished" podID="855c71f2-a025-49cb-84c8-0206b2ec7e08" containerID="ba22d52e2c88845b48608d81df18d063e6040d2d90a6f634187534bf7ca77329" exitCode=0 Feb 02 15:39:19 crc kubenswrapper[4733]: I0202 15:39:19.185481 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-nrbdq" event={"ID":"855c71f2-a025-49cb-84c8-0206b2ec7e08","Type":"ContainerDied","Data":"ba22d52e2c88845b48608d81df18d063e6040d2d90a6f634187534bf7ca77329"} Feb 02 15:39:19 crc kubenswrapper[4733]: I0202 15:39:19.185533 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-nrbdq" event={"ID":"855c71f2-a025-49cb-84c8-0206b2ec7e08","Type":"ContainerDied","Data":"6345b9f9c8b0745137ab98048d92f341bfd5f9f90ae157501f391a8143376b1b"} Feb 02 15:39:19 crc kubenswrapper[4733]: I0202 15:39:19.185553 4733 scope.go:117] "RemoveContainer" containerID="ba22d52e2c88845b48608d81df18d063e6040d2d90a6f634187534bf7ca77329" Feb 02 15:39:19 crc kubenswrapper[4733]: I0202 15:39:19.185662 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-nrbdq" Feb 02 15:39:19 crc kubenswrapper[4733]: I0202 15:39:19.187853 4733 generic.go:334] "Generic (PLEG): container finished" podID="2e4d2825-196e-4bf2-8f47-270645f2267b" containerID="99ee0047c1f019d60cd3d61bf905abf882cb05edbb2b3e61679da900a5465bc4" exitCode=0 Feb 02 15:39:19 crc kubenswrapper[4733]: I0202 15:39:19.187890 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-gzhj7" event={"ID":"2e4d2825-196e-4bf2-8f47-270645f2267b","Type":"ContainerDied","Data":"99ee0047c1f019d60cd3d61bf905abf882cb05edbb2b3e61679da900a5465bc4"} Feb 02 15:39:19 crc kubenswrapper[4733]: I0202 15:39:19.187923 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-gzhj7" event={"ID":"2e4d2825-196e-4bf2-8f47-270645f2267b","Type":"ContainerDied","Data":"fa50ca92d643ab3b4bda92b4db4e509f85a4a7b5d120380bb525eae0a60fd7b8"} Feb 02 15:39:19 crc kubenswrapper[4733]: I0202 15:39:19.187906 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-gzhj7" Feb 02 15:39:19 crc kubenswrapper[4733]: I0202 15:39:19.212931 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lc77f\" (UniqueName: \"kubernetes.io/projected/855c71f2-a025-49cb-84c8-0206b2ec7e08-kube-api-access-lc77f\") pod \"855c71f2-a025-49cb-84c8-0206b2ec7e08\" (UID: \"855c71f2-a025-49cb-84c8-0206b2ec7e08\") " Feb 02 15:39:19 crc kubenswrapper[4733]: I0202 15:39:19.217299 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-gzhj7"] Feb 02 15:39:19 crc kubenswrapper[4733]: I0202 15:39:19.222954 4733 scope.go:117] "RemoveContainer" containerID="ba22d52e2c88845b48608d81df18d063e6040d2d90a6f634187534bf7ca77329" Feb 02 15:39:19 crc kubenswrapper[4733]: E0202 15:39:19.223462 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba22d52e2c88845b48608d81df18d063e6040d2d90a6f634187534bf7ca77329\": container with ID starting with ba22d52e2c88845b48608d81df18d063e6040d2d90a6f634187534bf7ca77329 not found: ID does not exist" containerID="ba22d52e2c88845b48608d81df18d063e6040d2d90a6f634187534bf7ca77329" Feb 02 15:39:19 crc kubenswrapper[4733]: I0202 15:39:19.223503 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba22d52e2c88845b48608d81df18d063e6040d2d90a6f634187534bf7ca77329"} err="failed to get container status \"ba22d52e2c88845b48608d81df18d063e6040d2d90a6f634187534bf7ca77329\": rpc error: code = NotFound desc = could not find container \"ba22d52e2c88845b48608d81df18d063e6040d2d90a6f634187534bf7ca77329\": container with ID starting with ba22d52e2c88845b48608d81df18d063e6040d2d90a6f634187534bf7ca77329 not found: ID does not exist" Feb 02 15:39:19 crc kubenswrapper[4733]: I0202 15:39:19.223532 4733 scope.go:117] "RemoveContainer" containerID="99ee0047c1f019d60cd3d61bf905abf882cb05edbb2b3e61679da900a5465bc4" Feb 02 15:39:19 crc kubenswrapper[4733]: I0202 15:39:19.225314 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-gzhj7"] Feb 02 15:39:19 crc kubenswrapper[4733]: I0202 15:39:19.226893 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/855c71f2-a025-49cb-84c8-0206b2ec7e08-kube-api-access-lc77f" (OuterVolumeSpecName: "kube-api-access-lc77f") pod "855c71f2-a025-49cb-84c8-0206b2ec7e08" (UID: "855c71f2-a025-49cb-84c8-0206b2ec7e08"). InnerVolumeSpecName "kube-api-access-lc77f". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:39:19 crc kubenswrapper[4733]: I0202 15:39:19.238837 4733 scope.go:117] "RemoveContainer" containerID="99ee0047c1f019d60cd3d61bf905abf882cb05edbb2b3e61679da900a5465bc4" Feb 02 15:39:19 crc kubenswrapper[4733]: E0202 15:39:19.239277 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99ee0047c1f019d60cd3d61bf905abf882cb05edbb2b3e61679da900a5465bc4\": container with ID starting with 99ee0047c1f019d60cd3d61bf905abf882cb05edbb2b3e61679da900a5465bc4 not found: ID does not exist" containerID="99ee0047c1f019d60cd3d61bf905abf882cb05edbb2b3e61679da900a5465bc4" Feb 02 15:39:19 crc kubenswrapper[4733]: I0202 15:39:19.239313 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99ee0047c1f019d60cd3d61bf905abf882cb05edbb2b3e61679da900a5465bc4"} err="failed to get container status \"99ee0047c1f019d60cd3d61bf905abf882cb05edbb2b3e61679da900a5465bc4\": rpc error: code = NotFound desc = could not find container \"99ee0047c1f019d60cd3d61bf905abf882cb05edbb2b3e61679da900a5465bc4\": container with ID starting with 99ee0047c1f019d60cd3d61bf905abf882cb05edbb2b3e61679da900a5465bc4 not found: ID does not exist" Feb 02 15:39:19 crc kubenswrapper[4733]: I0202 15:39:19.264722 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e4d2825-196e-4bf2-8f47-270645f2267b" path="/var/lib/kubelet/pods/2e4d2825-196e-4bf2-8f47-270645f2267b/volumes" Feb 02 15:39:19 crc kubenswrapper[4733]: I0202 15:39:19.265286 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6" path="/var/lib/kubelet/pods/5cc5b7d3-4371-4f14-ae5b-712db7f1d2a6/volumes" Feb 02 15:39:19 crc kubenswrapper[4733]: I0202 15:39:19.314835 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lc77f\" (UniqueName: \"kubernetes.io/projected/855c71f2-a025-49cb-84c8-0206b2ec7e08-kube-api-access-lc77f\") on node \"crc\" DevicePath \"\"" Feb 02 15:39:19 crc kubenswrapper[4733]: I0202 15:39:19.351569 4733 scope.go:117] "RemoveContainer" containerID="02a2e1727447b5eada81717bc18a5e1fa01249afa350c42cb2d50d68ce2ee5fd" Feb 02 15:39:19 crc kubenswrapper[4733]: I0202 15:39:19.368604 4733 scope.go:117] "RemoveContainer" containerID="b7e3158461aaa2f50e0cf874860750cd5ffae4b2e58750199b1253c60286b0fc" Feb 02 15:39:19 crc kubenswrapper[4733]: I0202 15:39:19.392404 4733 scope.go:117] "RemoveContainer" containerID="a7d14acc7c8c90a61ee49da51a9d21cecb725ad53d1364bd144b14b64432aae5" Feb 02 15:39:19 crc kubenswrapper[4733]: I0202 15:39:19.409119 4733 scope.go:117] "RemoveContainer" containerID="8039566f2e14aac9a5310744a17130b451c216905c95273311f102f4d823d38e" Feb 02 15:39:19 crc kubenswrapper[4733]: I0202 15:39:19.436247 4733 scope.go:117] "RemoveContainer" containerID="5a1d2d8c4e19b20ac57cb8b31c117bc5dfc5e552e0082061c2b690e63e76f1c8" Feb 02 15:39:19 crc kubenswrapper[4733]: I0202 15:39:19.458014 4733 scope.go:117] "RemoveContainer" containerID="fdb7a4d2bece5b4300fa82ba55f8f2bb9a7e357f8b4066781a7982091f253779" Feb 02 15:39:19 crc kubenswrapper[4733]: I0202 15:39:19.487433 4733 scope.go:117] "RemoveContainer" containerID="05672c2c2a4105cf5a9fb1aca5de5bdb6d4df5b19893e7f79cc99cc627995e50" Feb 02 15:39:19 crc kubenswrapper[4733]: I0202 15:39:19.516508 4733 scope.go:117] "RemoveContainer" containerID="0eb75ac47796e3d56aedc56514b0b53601c1dcf4521b4e81651ca5e589242b0e" Feb 02 15:39:19 crc kubenswrapper[4733]: I0202 15:39:19.519493 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-nrbdq"] Feb 02 15:39:19 crc kubenswrapper[4733]: I0202 15:39:19.532177 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-nrbdq"] Feb 02 15:39:19 crc kubenswrapper[4733]: I0202 15:39:19.552204 4733 scope.go:117] "RemoveContainer" containerID="260c837b138cab3736b1ce776c2f8f709c073d15a9f64e6dd133c80772e67431" Feb 02 15:39:19 crc kubenswrapper[4733]: I0202 15:39:19.570736 4733 scope.go:117] "RemoveContainer" containerID="50bf7cb2672133cbe14cca992caced641278856ab52e9e82b6877bad6ca43b78" Feb 02 15:39:19 crc kubenswrapper[4733]: I0202 15:39:19.584504 4733 scope.go:117] "RemoveContainer" containerID="a3d0ca21093d66d012f1a132f486134fba8c350e499831e28e259fd9ad01f1ce" Feb 02 15:39:20 crc kubenswrapper[4733]: I0202 15:39:20.551092 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-controller-manager-b9d5f6fc5-nrtgx"] Feb 02 15:39:20 crc kubenswrapper[4733]: I0202 15:39:20.551399 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/infra-operator-controller-manager-b9d5f6fc5-nrtgx" podUID="7d80c76f-a56e-4c8b-9a02-73311129ea44" containerName="manager" containerID="cri-o://59e0e2d29d977f2cc61350355cd4e39bef9cd65be1e1ab4b2dfd1c4e70dbfa7a" gracePeriod=10 Feb 02 15:39:20 crc kubenswrapper[4733]: I0202 15:39:20.762061 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-index-552jg"] Feb 02 15:39:20 crc kubenswrapper[4733]: I0202 15:39:20.762688 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/infra-operator-index-552jg" podUID="bf01001b-fb23-4f09-8f88-38bce44a93ce" containerName="registry-server" containerID="cri-o://3368bbba9d3374e124d780779edd0479bbb6d0926fdb67ad3475b13e53ce6e61" gracePeriod=30 Feb 02 15:39:20 crc kubenswrapper[4733]: I0202 15:39:20.772259 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f75764trqm"] Feb 02 15:39:20 crc kubenswrapper[4733]: I0202 15:39:20.781238 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/d7c3b59ed6c2e571e21460d743e5fcd0c5f76cb7c446e474a3d05f75764trqm"] Feb 02 15:39:20 crc kubenswrapper[4733]: I0202 15:39:20.937291 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-b9d5f6fc5-nrtgx" Feb 02 15:39:21 crc kubenswrapper[4733]: I0202 15:39:21.037818 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7d80c76f-a56e-4c8b-9a02-73311129ea44-webhook-cert\") pod \"7d80c76f-a56e-4c8b-9a02-73311129ea44\" (UID: \"7d80c76f-a56e-4c8b-9a02-73311129ea44\") " Feb 02 15:39:21 crc kubenswrapper[4733]: I0202 15:39:21.037900 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nqpqq\" (UniqueName: \"kubernetes.io/projected/7d80c76f-a56e-4c8b-9a02-73311129ea44-kube-api-access-nqpqq\") pod \"7d80c76f-a56e-4c8b-9a02-73311129ea44\" (UID: \"7d80c76f-a56e-4c8b-9a02-73311129ea44\") " Feb 02 15:39:21 crc kubenswrapper[4733]: I0202 15:39:21.037924 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7d80c76f-a56e-4c8b-9a02-73311129ea44-apiservice-cert\") pod \"7d80c76f-a56e-4c8b-9a02-73311129ea44\" (UID: \"7d80c76f-a56e-4c8b-9a02-73311129ea44\") " Feb 02 15:39:21 crc kubenswrapper[4733]: I0202 15:39:21.043216 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d80c76f-a56e-4c8b-9a02-73311129ea44-kube-api-access-nqpqq" (OuterVolumeSpecName: "kube-api-access-nqpqq") pod "7d80c76f-a56e-4c8b-9a02-73311129ea44" (UID: "7d80c76f-a56e-4c8b-9a02-73311129ea44"). InnerVolumeSpecName "kube-api-access-nqpqq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:39:21 crc kubenswrapper[4733]: I0202 15:39:21.044359 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d80c76f-a56e-4c8b-9a02-73311129ea44-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "7d80c76f-a56e-4c8b-9a02-73311129ea44" (UID: "7d80c76f-a56e-4c8b-9a02-73311129ea44"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:39:21 crc kubenswrapper[4733]: I0202 15:39:21.044373 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d80c76f-a56e-4c8b-9a02-73311129ea44-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "7d80c76f-a56e-4c8b-9a02-73311129ea44" (UID: "7d80c76f-a56e-4c8b-9a02-73311129ea44"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:39:21 crc kubenswrapper[4733]: I0202 15:39:21.078896 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-552jg" Feb 02 15:39:21 crc kubenswrapper[4733]: I0202 15:39:21.139505 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nqpqq\" (UniqueName: \"kubernetes.io/projected/7d80c76f-a56e-4c8b-9a02-73311129ea44-kube-api-access-nqpqq\") on node \"crc\" DevicePath \"\"" Feb 02 15:39:21 crc kubenswrapper[4733]: I0202 15:39:21.139561 4733 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7d80c76f-a56e-4c8b-9a02-73311129ea44-apiservice-cert\") on node \"crc\" DevicePath \"\"" Feb 02 15:39:21 crc kubenswrapper[4733]: I0202 15:39:21.139583 4733 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7d80c76f-a56e-4c8b-9a02-73311129ea44-webhook-cert\") on node \"crc\" DevicePath \"\"" Feb 02 15:39:21 crc kubenswrapper[4733]: I0202 15:39:21.201724 4733 generic.go:334] "Generic (PLEG): container finished" podID="bf01001b-fb23-4f09-8f88-38bce44a93ce" containerID="3368bbba9d3374e124d780779edd0479bbb6d0926fdb67ad3475b13e53ce6e61" exitCode=0 Feb 02 15:39:21 crc kubenswrapper[4733]: I0202 15:39:21.201796 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-552jg" event={"ID":"bf01001b-fb23-4f09-8f88-38bce44a93ce","Type":"ContainerDied","Data":"3368bbba9d3374e124d780779edd0479bbb6d0926fdb67ad3475b13e53ce6e61"} Feb 02 15:39:21 crc kubenswrapper[4733]: I0202 15:39:21.201826 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-552jg" event={"ID":"bf01001b-fb23-4f09-8f88-38bce44a93ce","Type":"ContainerDied","Data":"72e9a7b68560d3215b10f90b7817deb03c361d6c35ac7081762db112eca7da16"} Feb 02 15:39:21 crc kubenswrapper[4733]: I0202 15:39:21.201825 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-552jg" Feb 02 15:39:21 crc kubenswrapper[4733]: I0202 15:39:21.201845 4733 scope.go:117] "RemoveContainer" containerID="3368bbba9d3374e124d780779edd0479bbb6d0926fdb67ad3475b13e53ce6e61" Feb 02 15:39:21 crc kubenswrapper[4733]: I0202 15:39:21.203776 4733 generic.go:334] "Generic (PLEG): container finished" podID="7d80c76f-a56e-4c8b-9a02-73311129ea44" containerID="59e0e2d29d977f2cc61350355cd4e39bef9cd65be1e1ab4b2dfd1c4e70dbfa7a" exitCode=0 Feb 02 15:39:21 crc kubenswrapper[4733]: I0202 15:39:21.203814 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-b9d5f6fc5-nrtgx" Feb 02 15:39:21 crc kubenswrapper[4733]: I0202 15:39:21.203823 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-b9d5f6fc5-nrtgx" event={"ID":"7d80c76f-a56e-4c8b-9a02-73311129ea44","Type":"ContainerDied","Data":"59e0e2d29d977f2cc61350355cd4e39bef9cd65be1e1ab4b2dfd1c4e70dbfa7a"} Feb 02 15:39:21 crc kubenswrapper[4733]: I0202 15:39:21.203852 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-b9d5f6fc5-nrtgx" event={"ID":"7d80c76f-a56e-4c8b-9a02-73311129ea44","Type":"ContainerDied","Data":"22507597bb0f5d49c018769e0899bc2cc2c58f459a851928b8c408cfd6ca98aa"} Feb 02 15:39:21 crc kubenswrapper[4733]: I0202 15:39:21.230972 4733 scope.go:117] "RemoveContainer" containerID="3368bbba9d3374e124d780779edd0479bbb6d0926fdb67ad3475b13e53ce6e61" Feb 02 15:39:21 crc kubenswrapper[4733]: E0202 15:39:21.231561 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3368bbba9d3374e124d780779edd0479bbb6d0926fdb67ad3475b13e53ce6e61\": container with ID starting with 3368bbba9d3374e124d780779edd0479bbb6d0926fdb67ad3475b13e53ce6e61 not found: ID does not exist" containerID="3368bbba9d3374e124d780779edd0479bbb6d0926fdb67ad3475b13e53ce6e61" Feb 02 15:39:21 crc kubenswrapper[4733]: I0202 15:39:21.231618 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3368bbba9d3374e124d780779edd0479bbb6d0926fdb67ad3475b13e53ce6e61"} err="failed to get container status \"3368bbba9d3374e124d780779edd0479bbb6d0926fdb67ad3475b13e53ce6e61\": rpc error: code = NotFound desc = could not find container \"3368bbba9d3374e124d780779edd0479bbb6d0926fdb67ad3475b13e53ce6e61\": container with ID starting with 3368bbba9d3374e124d780779edd0479bbb6d0926fdb67ad3475b13e53ce6e61 not found: ID does not exist" Feb 02 15:39:21 crc kubenswrapper[4733]: I0202 15:39:21.231646 4733 scope.go:117] "RemoveContainer" containerID="59e0e2d29d977f2cc61350355cd4e39bef9cd65be1e1ab4b2dfd1c4e70dbfa7a" Feb 02 15:39:21 crc kubenswrapper[4733]: I0202 15:39:21.240754 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mph99\" (UniqueName: \"kubernetes.io/projected/bf01001b-fb23-4f09-8f88-38bce44a93ce-kube-api-access-mph99\") pod \"bf01001b-fb23-4f09-8f88-38bce44a93ce\" (UID: \"bf01001b-fb23-4f09-8f88-38bce44a93ce\") " Feb 02 15:39:21 crc kubenswrapper[4733]: I0202 15:39:21.243261 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-controller-manager-b9d5f6fc5-nrtgx"] Feb 02 15:39:21 crc kubenswrapper[4733]: I0202 15:39:21.245417 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf01001b-fb23-4f09-8f88-38bce44a93ce-kube-api-access-mph99" (OuterVolumeSpecName: "kube-api-access-mph99") pod "bf01001b-fb23-4f09-8f88-38bce44a93ce" (UID: "bf01001b-fb23-4f09-8f88-38bce44a93ce"). InnerVolumeSpecName "kube-api-access-mph99". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:39:21 crc kubenswrapper[4733]: I0202 15:39:21.253133 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/infra-operator-controller-manager-b9d5f6fc5-nrtgx"] Feb 02 15:39:21 crc kubenswrapper[4733]: I0202 15:39:21.258578 4733 scope.go:117] "RemoveContainer" containerID="59e0e2d29d977f2cc61350355cd4e39bef9cd65be1e1ab4b2dfd1c4e70dbfa7a" Feb 02 15:39:21 crc kubenswrapper[4733]: E0202 15:39:21.260140 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59e0e2d29d977f2cc61350355cd4e39bef9cd65be1e1ab4b2dfd1c4e70dbfa7a\": container with ID starting with 59e0e2d29d977f2cc61350355cd4e39bef9cd65be1e1ab4b2dfd1c4e70dbfa7a not found: ID does not exist" containerID="59e0e2d29d977f2cc61350355cd4e39bef9cd65be1e1ab4b2dfd1c4e70dbfa7a" Feb 02 15:39:21 crc kubenswrapper[4733]: I0202 15:39:21.260221 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59e0e2d29d977f2cc61350355cd4e39bef9cd65be1e1ab4b2dfd1c4e70dbfa7a"} err="failed to get container status \"59e0e2d29d977f2cc61350355cd4e39bef9cd65be1e1ab4b2dfd1c4e70dbfa7a\": rpc error: code = NotFound desc = could not find container \"59e0e2d29d977f2cc61350355cd4e39bef9cd65be1e1ab4b2dfd1c4e70dbfa7a\": container with ID starting with 59e0e2d29d977f2cc61350355cd4e39bef9cd65be1e1ab4b2dfd1c4e70dbfa7a not found: ID does not exist" Feb 02 15:39:21 crc kubenswrapper[4733]: I0202 15:39:21.265488 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d80c76f-a56e-4c8b-9a02-73311129ea44" path="/var/lib/kubelet/pods/7d80c76f-a56e-4c8b-9a02-73311129ea44/volumes" Feb 02 15:39:21 crc kubenswrapper[4733]: I0202 15:39:21.266259 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="855c71f2-a025-49cb-84c8-0206b2ec7e08" path="/var/lib/kubelet/pods/855c71f2-a025-49cb-84c8-0206b2ec7e08/volumes" Feb 02 15:39:21 crc kubenswrapper[4733]: I0202 15:39:21.267017 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f107fa8c-f81f-4d3f-8a14-e718942491b8" path="/var/lib/kubelet/pods/f107fa8c-f81f-4d3f-8a14-e718942491b8/volumes" Feb 02 15:39:21 crc kubenswrapper[4733]: I0202 15:39:21.342002 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mph99\" (UniqueName: \"kubernetes.io/projected/bf01001b-fb23-4f09-8f88-38bce44a93ce-kube-api-access-mph99\") on node \"crc\" DevicePath \"\"" Feb 02 15:39:21 crc kubenswrapper[4733]: I0202 15:39:21.520290 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-index-552jg"] Feb 02 15:39:21 crc kubenswrapper[4733]: I0202 15:39:21.526027 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/infra-operator-index-552jg"] Feb 02 15:39:22 crc kubenswrapper[4733]: I0202 15:39:22.083295 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6984697d4d-5mvzr"] Feb 02 15:39:22 crc kubenswrapper[4733]: I0202 15:39:22.083998 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/mariadb-operator-controller-manager-6984697d4d-5mvzr" podUID="ccaf63aa-d3fe-492e-8f45-0502e4f66cd3" containerName="manager" containerID="cri-o://fc36ca98de458bfe957316ab2c04717dbf8f5a235e9cca3517345554e008d530" gracePeriod=10 Feb 02 15:39:22 crc kubenswrapper[4733]: I0202 15:39:22.300430 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-bmfgc"] Feb 02 15:39:22 crc kubenswrapper[4733]: I0202 15:39:22.307760 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/mariadb-operator-index-bmfgc" podUID="1904d4f7-5ff6-4b7f-ba4a-5bf63ac6f95a" containerName="registry-server" containerID="cri-o://b65578e7d7a4500ed6660821b27c40d43d0cc1fe5e3a180af3050acea260f47b" gracePeriod=30 Feb 02 15:39:22 crc kubenswrapper[4733]: I0202 15:39:22.334528 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40qpth5"] Feb 02 15:39:22 crc kubenswrapper[4733]: I0202 15:39:22.339280 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/f5f7435db1a968bc2e4b919cf4f5a8f6719d9ac995e6b095f5b2e84f40qpth5"] Feb 02 15:39:22 crc kubenswrapper[4733]: I0202 15:39:22.489215 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6984697d4d-5mvzr" Feb 02 15:39:22 crc kubenswrapper[4733]: I0202 15:39:22.596326 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ccaf63aa-d3fe-492e-8f45-0502e4f66cd3-apiservice-cert\") pod \"ccaf63aa-d3fe-492e-8f45-0502e4f66cd3\" (UID: \"ccaf63aa-d3fe-492e-8f45-0502e4f66cd3\") " Feb 02 15:39:22 crc kubenswrapper[4733]: I0202 15:39:22.596776 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zmjsm\" (UniqueName: \"kubernetes.io/projected/ccaf63aa-d3fe-492e-8f45-0502e4f66cd3-kube-api-access-zmjsm\") pod \"ccaf63aa-d3fe-492e-8f45-0502e4f66cd3\" (UID: \"ccaf63aa-d3fe-492e-8f45-0502e4f66cd3\") " Feb 02 15:39:22 crc kubenswrapper[4733]: I0202 15:39:22.596863 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ccaf63aa-d3fe-492e-8f45-0502e4f66cd3-webhook-cert\") pod \"ccaf63aa-d3fe-492e-8f45-0502e4f66cd3\" (UID: \"ccaf63aa-d3fe-492e-8f45-0502e4f66cd3\") " Feb 02 15:39:22 crc kubenswrapper[4733]: I0202 15:39:22.605182 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccaf63aa-d3fe-492e-8f45-0502e4f66cd3-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "ccaf63aa-d3fe-492e-8f45-0502e4f66cd3" (UID: "ccaf63aa-d3fe-492e-8f45-0502e4f66cd3"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:39:22 crc kubenswrapper[4733]: I0202 15:39:22.610124 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccaf63aa-d3fe-492e-8f45-0502e4f66cd3-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "ccaf63aa-d3fe-492e-8f45-0502e4f66cd3" (UID: "ccaf63aa-d3fe-492e-8f45-0502e4f66cd3"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:39:22 crc kubenswrapper[4733]: I0202 15:39:22.619878 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccaf63aa-d3fe-492e-8f45-0502e4f66cd3-kube-api-access-zmjsm" (OuterVolumeSpecName: "kube-api-access-zmjsm") pod "ccaf63aa-d3fe-492e-8f45-0502e4f66cd3" (UID: "ccaf63aa-d3fe-492e-8f45-0502e4f66cd3"). InnerVolumeSpecName "kube-api-access-zmjsm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:39:22 crc kubenswrapper[4733]: I0202 15:39:22.698933 4733 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ccaf63aa-d3fe-492e-8f45-0502e4f66cd3-webhook-cert\") on node \"crc\" DevicePath \"\"" Feb 02 15:39:22 crc kubenswrapper[4733]: I0202 15:39:22.698969 4733 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ccaf63aa-d3fe-492e-8f45-0502e4f66cd3-apiservice-cert\") on node \"crc\" DevicePath \"\"" Feb 02 15:39:22 crc kubenswrapper[4733]: I0202 15:39:22.698983 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zmjsm\" (UniqueName: \"kubernetes.io/projected/ccaf63aa-d3fe-492e-8f45-0502e4f66cd3-kube-api-access-zmjsm\") on node \"crc\" DevicePath \"\"" Feb 02 15:39:22 crc kubenswrapper[4733]: I0202 15:39:22.776719 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-bmfgc" Feb 02 15:39:22 crc kubenswrapper[4733]: I0202 15:39:22.901310 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p52nr\" (UniqueName: \"kubernetes.io/projected/1904d4f7-5ff6-4b7f-ba4a-5bf63ac6f95a-kube-api-access-p52nr\") pod \"1904d4f7-5ff6-4b7f-ba4a-5bf63ac6f95a\" (UID: \"1904d4f7-5ff6-4b7f-ba4a-5bf63ac6f95a\") " Feb 02 15:39:22 crc kubenswrapper[4733]: I0202 15:39:22.906579 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1904d4f7-5ff6-4b7f-ba4a-5bf63ac6f95a-kube-api-access-p52nr" (OuterVolumeSpecName: "kube-api-access-p52nr") pod "1904d4f7-5ff6-4b7f-ba4a-5bf63ac6f95a" (UID: "1904d4f7-5ff6-4b7f-ba4a-5bf63ac6f95a"). InnerVolumeSpecName "kube-api-access-p52nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:39:23 crc kubenswrapper[4733]: I0202 15:39:23.003187 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p52nr\" (UniqueName: \"kubernetes.io/projected/1904d4f7-5ff6-4b7f-ba4a-5bf63ac6f95a-kube-api-access-p52nr\") on node \"crc\" DevicePath \"\"" Feb 02 15:39:23 crc kubenswrapper[4733]: I0202 15:39:23.222750 4733 generic.go:334] "Generic (PLEG): container finished" podID="1904d4f7-5ff6-4b7f-ba4a-5bf63ac6f95a" containerID="b65578e7d7a4500ed6660821b27c40d43d0cc1fe5e3a180af3050acea260f47b" exitCode=0 Feb 02 15:39:23 crc kubenswrapper[4733]: I0202 15:39:23.222819 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-bmfgc" Feb 02 15:39:23 crc kubenswrapper[4733]: I0202 15:39:23.222890 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-bmfgc" event={"ID":"1904d4f7-5ff6-4b7f-ba4a-5bf63ac6f95a","Type":"ContainerDied","Data":"b65578e7d7a4500ed6660821b27c40d43d0cc1fe5e3a180af3050acea260f47b"} Feb 02 15:39:23 crc kubenswrapper[4733]: I0202 15:39:23.222977 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-bmfgc" event={"ID":"1904d4f7-5ff6-4b7f-ba4a-5bf63ac6f95a","Type":"ContainerDied","Data":"e01105b3fbedb424c8f2df21c648539b996f8a48a3e431902bde68d2aeb1e749"} Feb 02 15:39:23 crc kubenswrapper[4733]: I0202 15:39:23.223015 4733 scope.go:117] "RemoveContainer" containerID="b65578e7d7a4500ed6660821b27c40d43d0cc1fe5e3a180af3050acea260f47b" Feb 02 15:39:23 crc kubenswrapper[4733]: I0202 15:39:23.234592 4733 generic.go:334] "Generic (PLEG): container finished" podID="ccaf63aa-d3fe-492e-8f45-0502e4f66cd3" containerID="fc36ca98de458bfe957316ab2c04717dbf8f5a235e9cca3517345554e008d530" exitCode=0 Feb 02 15:39:23 crc kubenswrapper[4733]: I0202 15:39:23.234658 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6984697d4d-5mvzr" event={"ID":"ccaf63aa-d3fe-492e-8f45-0502e4f66cd3","Type":"ContainerDied","Data":"fc36ca98de458bfe957316ab2c04717dbf8f5a235e9cca3517345554e008d530"} Feb 02 15:39:23 crc kubenswrapper[4733]: I0202 15:39:23.234697 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6984697d4d-5mvzr" event={"ID":"ccaf63aa-d3fe-492e-8f45-0502e4f66cd3","Type":"ContainerDied","Data":"88ccd8fde787861154d902dde44cb3bcb6c57e268ef0ec80c9d4a70d9761f163"} Feb 02 15:39:23 crc kubenswrapper[4733]: I0202 15:39:23.234787 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6984697d4d-5mvzr" Feb 02 15:39:23 crc kubenswrapper[4733]: I0202 15:39:23.256808 4733 scope.go:117] "RemoveContainer" containerID="b65578e7d7a4500ed6660821b27c40d43d0cc1fe5e3a180af3050acea260f47b" Feb 02 15:39:23 crc kubenswrapper[4733]: E0202 15:39:23.257560 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b65578e7d7a4500ed6660821b27c40d43d0cc1fe5e3a180af3050acea260f47b\": container with ID starting with b65578e7d7a4500ed6660821b27c40d43d0cc1fe5e3a180af3050acea260f47b not found: ID does not exist" containerID="b65578e7d7a4500ed6660821b27c40d43d0cc1fe5e3a180af3050acea260f47b" Feb 02 15:39:23 crc kubenswrapper[4733]: I0202 15:39:23.257657 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b65578e7d7a4500ed6660821b27c40d43d0cc1fe5e3a180af3050acea260f47b"} err="failed to get container status \"b65578e7d7a4500ed6660821b27c40d43d0cc1fe5e3a180af3050acea260f47b\": rpc error: code = NotFound desc = could not find container \"b65578e7d7a4500ed6660821b27c40d43d0cc1fe5e3a180af3050acea260f47b\": container with ID starting with b65578e7d7a4500ed6660821b27c40d43d0cc1fe5e3a180af3050acea260f47b not found: ID does not exist" Feb 02 15:39:23 crc kubenswrapper[4733]: I0202 15:39:23.257697 4733 scope.go:117] "RemoveContainer" containerID="fc36ca98de458bfe957316ab2c04717dbf8f5a235e9cca3517345554e008d530" Feb 02 15:39:23 crc kubenswrapper[4733]: I0202 15:39:23.269840 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40e65f33-ed0b-4d65-8a08-4a4ee970ea0f" path="/var/lib/kubelet/pods/40e65f33-ed0b-4d65-8a08-4a4ee970ea0f/volumes" Feb 02 15:39:23 crc kubenswrapper[4733]: I0202 15:39:23.276876 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf01001b-fb23-4f09-8f88-38bce44a93ce" path="/var/lib/kubelet/pods/bf01001b-fb23-4f09-8f88-38bce44a93ce/volumes" Feb 02 15:39:23 crc kubenswrapper[4733]: I0202 15:39:23.278569 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-bmfgc"] Feb 02 15:39:23 crc kubenswrapper[4733]: I0202 15:39:23.288460 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/mariadb-operator-index-bmfgc"] Feb 02 15:39:23 crc kubenswrapper[4733]: I0202 15:39:23.293391 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6984697d4d-5mvzr"] Feb 02 15:39:23 crc kubenswrapper[4733]: I0202 15:39:23.300057 4733 scope.go:117] "RemoveContainer" containerID="fc36ca98de458bfe957316ab2c04717dbf8f5a235e9cca3517345554e008d530" Feb 02 15:39:23 crc kubenswrapper[4733]: E0202 15:39:23.300703 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc36ca98de458bfe957316ab2c04717dbf8f5a235e9cca3517345554e008d530\": container with ID starting with fc36ca98de458bfe957316ab2c04717dbf8f5a235e9cca3517345554e008d530 not found: ID does not exist" containerID="fc36ca98de458bfe957316ab2c04717dbf8f5a235e9cca3517345554e008d530" Feb 02 15:39:23 crc kubenswrapper[4733]: I0202 15:39:23.300772 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc36ca98de458bfe957316ab2c04717dbf8f5a235e9cca3517345554e008d530"} err="failed to get container status \"fc36ca98de458bfe957316ab2c04717dbf8f5a235e9cca3517345554e008d530\": rpc error: code = NotFound desc = could not find container \"fc36ca98de458bfe957316ab2c04717dbf8f5a235e9cca3517345554e008d530\": container with ID starting with fc36ca98de458bfe957316ab2c04717dbf8f5a235e9cca3517345554e008d530 not found: ID does not exist" Feb 02 15:39:23 crc kubenswrapper[4733]: I0202 15:39:23.308476 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6984697d4d-5mvzr"] Feb 02 15:39:25 crc kubenswrapper[4733]: I0202 15:39:25.268549 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1904d4f7-5ff6-4b7f-ba4a-5bf63ac6f95a" path="/var/lib/kubelet/pods/1904d4f7-5ff6-4b7f-ba4a-5bf63ac6f95a/volumes" Feb 02 15:39:25 crc kubenswrapper[4733]: I0202 15:39:25.269992 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccaf63aa-d3fe-492e-8f45-0502e4f66cd3" path="/var/lib/kubelet/pods/ccaf63aa-d3fe-492e-8f45-0502e4f66cd3/volumes" Feb 02 15:39:26 crc kubenswrapper[4733]: I0202 15:39:26.680919 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-t2vw2"] Feb 02 15:39:26 crc kubenswrapper[4733]: E0202 15:39:26.681592 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="484de6bc-40c4-4159-a4f5-9cf345357b40" containerName="extract-content" Feb 02 15:39:26 crc kubenswrapper[4733]: I0202 15:39:26.681611 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="484de6bc-40c4-4159-a4f5-9cf345357b40" containerName="extract-content" Feb 02 15:39:26 crc kubenswrapper[4733]: E0202 15:39:26.681662 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d63f3e93-f45f-4f7d-a29b-2eaae536fa5d" containerName="mysql-bootstrap" Feb 02 15:39:26 crc kubenswrapper[4733]: I0202 15:39:26.681674 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="d63f3e93-f45f-4f7d-a29b-2eaae536fa5d" containerName="mysql-bootstrap" Feb 02 15:39:26 crc kubenswrapper[4733]: E0202 15:39:26.682015 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccaf63aa-d3fe-492e-8f45-0502e4f66cd3" containerName="manager" Feb 02 15:39:26 crc kubenswrapper[4733]: I0202 15:39:26.682026 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccaf63aa-d3fe-492e-8f45-0502e4f66cd3" containerName="manager" Feb 02 15:39:26 crc kubenswrapper[4733]: E0202 15:39:26.682043 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f45e90dd-ab9d-4c27-8449-98cf52f48c87" containerName="rabbitmq" Feb 02 15:39:26 crc kubenswrapper[4733]: I0202 15:39:26.682054 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f45e90dd-ab9d-4c27-8449-98cf52f48c87" containerName="rabbitmq" Feb 02 15:39:26 crc kubenswrapper[4733]: E0202 15:39:26.682073 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f177ebf9-728b-4ba2-b9b7-6c676f9d014a" containerName="mysql-bootstrap" Feb 02 15:39:26 crc kubenswrapper[4733]: I0202 15:39:26.682084 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f177ebf9-728b-4ba2-b9b7-6c676f9d014a" containerName="mysql-bootstrap" Feb 02 15:39:26 crc kubenswrapper[4733]: E0202 15:39:26.682103 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d80c76f-a56e-4c8b-9a02-73311129ea44" containerName="manager" Feb 02 15:39:26 crc kubenswrapper[4733]: I0202 15:39:26.682113 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d80c76f-a56e-4c8b-9a02-73311129ea44" containerName="manager" Feb 02 15:39:26 crc kubenswrapper[4733]: E0202 15:39:26.682129 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f45e90dd-ab9d-4c27-8449-98cf52f48c87" containerName="setup-container" Feb 02 15:39:26 crc kubenswrapper[4733]: I0202 15:39:26.682140 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f45e90dd-ab9d-4c27-8449-98cf52f48c87" containerName="setup-container" Feb 02 15:39:26 crc kubenswrapper[4733]: E0202 15:39:26.682157 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dffed5a-a57d-4710-909f-799b47ff68ef" containerName="manila-service-cleanup-n5b5h655" Feb 02 15:39:26 crc kubenswrapper[4733]: I0202 15:39:26.682190 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dffed5a-a57d-4710-909f-799b47ff68ef" containerName="manila-service-cleanup-n5b5h655" Feb 02 15:39:26 crc kubenswrapper[4733]: E0202 15:39:26.682208 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d63f3e93-f45f-4f7d-a29b-2eaae536fa5d" containerName="galera" Feb 02 15:39:26 crc kubenswrapper[4733]: I0202 15:39:26.682218 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="d63f3e93-f45f-4f7d-a29b-2eaae536fa5d" containerName="galera" Feb 02 15:39:26 crc kubenswrapper[4733]: E0202 15:39:26.682236 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79f9eeab-dccd-4404-8179-fc3f776d8b22" containerName="mariadb-account-delete" Feb 02 15:39:26 crc kubenswrapper[4733]: I0202 15:39:26.682246 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="79f9eeab-dccd-4404-8179-fc3f776d8b22" containerName="mariadb-account-delete" Feb 02 15:39:26 crc kubenswrapper[4733]: E0202 15:39:26.682261 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf01001b-fb23-4f09-8f88-38bce44a93ce" containerName="registry-server" Feb 02 15:39:26 crc kubenswrapper[4733]: I0202 15:39:26.682271 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf01001b-fb23-4f09-8f88-38bce44a93ce" containerName="registry-server" Feb 02 15:39:26 crc kubenswrapper[4733]: E0202 15:39:26.682286 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e4d2825-196e-4bf2-8f47-270645f2267b" containerName="operator" Feb 02 15:39:26 crc kubenswrapper[4733]: I0202 15:39:26.682296 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e4d2825-196e-4bf2-8f47-270645f2267b" containerName="operator" Feb 02 15:39:26 crc kubenswrapper[4733]: E0202 15:39:26.682315 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c71eca7f-f211-40bc-9116-f3c246c46e54" containerName="ceph" Feb 02 15:39:26 crc kubenswrapper[4733]: I0202 15:39:26.682324 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="c71eca7f-f211-40bc-9116-f3c246c46e54" containerName="ceph" Feb 02 15:39:26 crc kubenswrapper[4733]: E0202 15:39:26.682335 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="484de6bc-40c4-4159-a4f5-9cf345357b40" containerName="registry-server" Feb 02 15:39:26 crc kubenswrapper[4733]: I0202 15:39:26.682344 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="484de6bc-40c4-4159-a4f5-9cf345357b40" containerName="registry-server" Feb 02 15:39:26 crc kubenswrapper[4733]: E0202 15:39:26.682356 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="855c71f2-a025-49cb-84c8-0206b2ec7e08" containerName="registry-server" Feb 02 15:39:26 crc kubenswrapper[4733]: I0202 15:39:26.682367 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="855c71f2-a025-49cb-84c8-0206b2ec7e08" containerName="registry-server" Feb 02 15:39:26 crc kubenswrapper[4733]: E0202 15:39:26.682380 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1aaf01fd-0114-41a6-af39-bd1a10deabe9" containerName="manager" Feb 02 15:39:26 crc kubenswrapper[4733]: I0202 15:39:26.682395 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="1aaf01fd-0114-41a6-af39-bd1a10deabe9" containerName="manager" Feb 02 15:39:26 crc kubenswrapper[4733]: E0202 15:39:26.682412 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1904d4f7-5ff6-4b7f-ba4a-5bf63ac6f95a" containerName="registry-server" Feb 02 15:39:26 crc kubenswrapper[4733]: I0202 15:39:26.682422 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="1904d4f7-5ff6-4b7f-ba4a-5bf63ac6f95a" containerName="registry-server" Feb 02 15:39:26 crc kubenswrapper[4733]: E0202 15:39:26.682436 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79f9eeab-dccd-4404-8179-fc3f776d8b22" containerName="mariadb-account-delete" Feb 02 15:39:26 crc kubenswrapper[4733]: I0202 15:39:26.682444 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="79f9eeab-dccd-4404-8179-fc3f776d8b22" containerName="mariadb-account-delete" Feb 02 15:39:26 crc kubenswrapper[4733]: E0202 15:39:26.682497 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f177ebf9-728b-4ba2-b9b7-6c676f9d014a" containerName="galera" Feb 02 15:39:26 crc kubenswrapper[4733]: I0202 15:39:26.682509 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="f177ebf9-728b-4ba2-b9b7-6c676f9d014a" containerName="galera" Feb 02 15:39:26 crc kubenswrapper[4733]: E0202 15:39:26.682525 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0147e51-446d-4a71-aaaa-e852258495d4" containerName="registry-server" Feb 02 15:39:26 crc kubenswrapper[4733]: I0202 15:39:26.682533 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0147e51-446d-4a71-aaaa-e852258495d4" containerName="registry-server" Feb 02 15:39:26 crc kubenswrapper[4733]: E0202 15:39:26.682543 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="484de6bc-40c4-4159-a4f5-9cf345357b40" containerName="extract-utilities" Feb 02 15:39:26 crc kubenswrapper[4733]: I0202 15:39:26.682552 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="484de6bc-40c4-4159-a4f5-9cf345357b40" containerName="extract-utilities" Feb 02 15:39:26 crc kubenswrapper[4733]: I0202 15:39:26.682721 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0147e51-446d-4a71-aaaa-e852258495d4" containerName="registry-server" Feb 02 15:39:26 crc kubenswrapper[4733]: I0202 15:39:26.682743 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="c71eca7f-f211-40bc-9116-f3c246c46e54" containerName="ceph" Feb 02 15:39:26 crc kubenswrapper[4733]: I0202 15:39:26.682759 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf01001b-fb23-4f09-8f88-38bce44a93ce" containerName="registry-server" Feb 02 15:39:26 crc kubenswrapper[4733]: I0202 15:39:26.682771 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f177ebf9-728b-4ba2-b9b7-6c676f9d014a" containerName="galera" Feb 02 15:39:26 crc kubenswrapper[4733]: I0202 15:39:26.682784 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="d63f3e93-f45f-4f7d-a29b-2eaae536fa5d" containerName="galera" Feb 02 15:39:26 crc kubenswrapper[4733]: I0202 15:39:26.682801 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="855c71f2-a025-49cb-84c8-0206b2ec7e08" containerName="registry-server" Feb 02 15:39:26 crc kubenswrapper[4733]: I0202 15:39:26.682813 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="1904d4f7-5ff6-4b7f-ba4a-5bf63ac6f95a" containerName="registry-server" Feb 02 15:39:26 crc kubenswrapper[4733]: I0202 15:39:26.682826 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="1aaf01fd-0114-41a6-af39-bd1a10deabe9" containerName="manager" Feb 02 15:39:26 crc kubenswrapper[4733]: I0202 15:39:26.682840 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccaf63aa-d3fe-492e-8f45-0502e4f66cd3" containerName="manager" Feb 02 15:39:26 crc kubenswrapper[4733]: I0202 15:39:26.682857 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="79f9eeab-dccd-4404-8179-fc3f776d8b22" containerName="mariadb-account-delete" Feb 02 15:39:26 crc kubenswrapper[4733]: I0202 15:39:26.682870 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="f45e90dd-ab9d-4c27-8449-98cf52f48c87" containerName="rabbitmq" Feb 02 15:39:26 crc kubenswrapper[4733]: I0202 15:39:26.682879 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="79f9eeab-dccd-4404-8179-fc3f776d8b22" containerName="mariadb-account-delete" Feb 02 15:39:26 crc kubenswrapper[4733]: I0202 15:39:26.682894 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="484de6bc-40c4-4159-a4f5-9cf345357b40" containerName="registry-server" Feb 02 15:39:26 crc kubenswrapper[4733]: I0202 15:39:26.682908 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d80c76f-a56e-4c8b-9a02-73311129ea44" containerName="manager" Feb 02 15:39:26 crc kubenswrapper[4733]: I0202 15:39:26.682919 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="9dffed5a-a57d-4710-909f-799b47ff68ef" containerName="manila-service-cleanup-n5b5h655" Feb 02 15:39:26 crc kubenswrapper[4733]: I0202 15:39:26.682945 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e4d2825-196e-4bf2-8f47-270645f2267b" containerName="operator" Feb 02 15:39:26 crc kubenswrapper[4733]: I0202 15:39:26.684380 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t2vw2" Feb 02 15:39:26 crc kubenswrapper[4733]: I0202 15:39:26.704245 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-t2vw2"] Feb 02 15:39:26 crc kubenswrapper[4733]: I0202 15:39:26.856808 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ef62ffe-f518-4f9e-86fb-8be7f18ef773-catalog-content\") pod \"redhat-marketplace-t2vw2\" (UID: \"8ef62ffe-f518-4f9e-86fb-8be7f18ef773\") " pod="openshift-marketplace/redhat-marketplace-t2vw2" Feb 02 15:39:26 crc kubenswrapper[4733]: I0202 15:39:26.857190 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cx9hp\" (UniqueName: \"kubernetes.io/projected/8ef62ffe-f518-4f9e-86fb-8be7f18ef773-kube-api-access-cx9hp\") pod \"redhat-marketplace-t2vw2\" (UID: \"8ef62ffe-f518-4f9e-86fb-8be7f18ef773\") " pod="openshift-marketplace/redhat-marketplace-t2vw2" Feb 02 15:39:26 crc kubenswrapper[4733]: I0202 15:39:26.857467 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ef62ffe-f518-4f9e-86fb-8be7f18ef773-utilities\") pod \"redhat-marketplace-t2vw2\" (UID: \"8ef62ffe-f518-4f9e-86fb-8be7f18ef773\") " pod="openshift-marketplace/redhat-marketplace-t2vw2" Feb 02 15:39:26 crc kubenswrapper[4733]: I0202 15:39:26.958716 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ef62ffe-f518-4f9e-86fb-8be7f18ef773-utilities\") pod \"redhat-marketplace-t2vw2\" (UID: \"8ef62ffe-f518-4f9e-86fb-8be7f18ef773\") " pod="openshift-marketplace/redhat-marketplace-t2vw2" Feb 02 15:39:26 crc kubenswrapper[4733]: I0202 15:39:26.958869 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ef62ffe-f518-4f9e-86fb-8be7f18ef773-catalog-content\") pod \"redhat-marketplace-t2vw2\" (UID: \"8ef62ffe-f518-4f9e-86fb-8be7f18ef773\") " pod="openshift-marketplace/redhat-marketplace-t2vw2" Feb 02 15:39:26 crc kubenswrapper[4733]: I0202 15:39:26.958926 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cx9hp\" (UniqueName: \"kubernetes.io/projected/8ef62ffe-f518-4f9e-86fb-8be7f18ef773-kube-api-access-cx9hp\") pod \"redhat-marketplace-t2vw2\" (UID: \"8ef62ffe-f518-4f9e-86fb-8be7f18ef773\") " pod="openshift-marketplace/redhat-marketplace-t2vw2" Feb 02 15:39:26 crc kubenswrapper[4733]: I0202 15:39:26.959718 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ef62ffe-f518-4f9e-86fb-8be7f18ef773-utilities\") pod \"redhat-marketplace-t2vw2\" (UID: \"8ef62ffe-f518-4f9e-86fb-8be7f18ef773\") " pod="openshift-marketplace/redhat-marketplace-t2vw2" Feb 02 15:39:26 crc kubenswrapper[4733]: I0202 15:39:26.959726 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ef62ffe-f518-4f9e-86fb-8be7f18ef773-catalog-content\") pod \"redhat-marketplace-t2vw2\" (UID: \"8ef62ffe-f518-4f9e-86fb-8be7f18ef773\") " pod="openshift-marketplace/redhat-marketplace-t2vw2" Feb 02 15:39:26 crc kubenswrapper[4733]: I0202 15:39:26.985355 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cx9hp\" (UniqueName: \"kubernetes.io/projected/8ef62ffe-f518-4f9e-86fb-8be7f18ef773-kube-api-access-cx9hp\") pod \"redhat-marketplace-t2vw2\" (UID: \"8ef62ffe-f518-4f9e-86fb-8be7f18ef773\") " pod="openshift-marketplace/redhat-marketplace-t2vw2" Feb 02 15:39:27 crc kubenswrapper[4733]: I0202 15:39:27.017456 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t2vw2" Feb 02 15:39:27 crc kubenswrapper[4733]: I0202 15:39:27.469967 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-t2vw2"] Feb 02 15:39:28 crc kubenswrapper[4733]: I0202 15:39:28.291883 4733 generic.go:334] "Generic (PLEG): container finished" podID="8ef62ffe-f518-4f9e-86fb-8be7f18ef773" containerID="df3f4e67398d7e7311ebdc12020fee8953b18a916f1565c8f7f84ec249a5a11b" exitCode=0 Feb 02 15:39:28 crc kubenswrapper[4733]: I0202 15:39:28.291939 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t2vw2" event={"ID":"8ef62ffe-f518-4f9e-86fb-8be7f18ef773","Type":"ContainerDied","Data":"df3f4e67398d7e7311ebdc12020fee8953b18a916f1565c8f7f84ec249a5a11b"} Feb 02 15:39:28 crc kubenswrapper[4733]: I0202 15:39:28.292430 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t2vw2" event={"ID":"8ef62ffe-f518-4f9e-86fb-8be7f18ef773","Type":"ContainerStarted","Data":"f3706c857d148eb9d0f3c15115295714fd60d41c38777689a81b0f7e4b80e0af"} Feb 02 15:39:29 crc kubenswrapper[4733]: I0202 15:39:29.306210 4733 generic.go:334] "Generic (PLEG): container finished" podID="8ef62ffe-f518-4f9e-86fb-8be7f18ef773" containerID="a7766ebd44e9718fcce4b6ba02ff3771ae6a21da6f35443e6c802a2360018967" exitCode=0 Feb 02 15:39:29 crc kubenswrapper[4733]: I0202 15:39:29.306301 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t2vw2" event={"ID":"8ef62ffe-f518-4f9e-86fb-8be7f18ef773","Type":"ContainerDied","Data":"a7766ebd44e9718fcce4b6ba02ff3771ae6a21da6f35443e6c802a2360018967"} Feb 02 15:39:30 crc kubenswrapper[4733]: I0202 15:39:30.317960 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t2vw2" event={"ID":"8ef62ffe-f518-4f9e-86fb-8be7f18ef773","Type":"ContainerStarted","Data":"3d2d5d7b87c0a19fe20b26558588ec17d316aa916aa594995a31509e4d76c00b"} Feb 02 15:39:30 crc kubenswrapper[4733]: I0202 15:39:30.351904 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-t2vw2" podStartSLOduration=2.590887805 podStartE2EDuration="4.351886333s" podCreationTimestamp="2026-02-02 15:39:26 +0000 UTC" firstStartedPulling="2026-02-02 15:39:28.293832161 +0000 UTC m=+1451.745293539" lastFinishedPulling="2026-02-02 15:39:30.054830699 +0000 UTC m=+1453.506292067" observedRunningTime="2026-02-02 15:39:30.34679712 +0000 UTC m=+1453.798258488" watchObservedRunningTime="2026-02-02 15:39:30.351886333 +0000 UTC m=+1453.803347701" Feb 02 15:39:34 crc kubenswrapper[4733]: I0202 15:39:34.988383 4733 patch_prober.go:28] interesting pod/machine-config-daemon-zgsv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 15:39:34 crc kubenswrapper[4733]: I0202 15:39:34.989012 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 15:39:37 crc kubenswrapper[4733]: I0202 15:39:37.018439 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-t2vw2" Feb 02 15:39:37 crc kubenswrapper[4733]: I0202 15:39:37.018514 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-t2vw2" Feb 02 15:39:37 crc kubenswrapper[4733]: I0202 15:39:37.096299 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-t2vw2" Feb 02 15:39:37 crc kubenswrapper[4733]: I0202 15:39:37.447253 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-t2vw2" Feb 02 15:39:37 crc kubenswrapper[4733]: I0202 15:39:37.513958 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-t2vw2"] Feb 02 15:39:39 crc kubenswrapper[4733]: I0202 15:39:39.399809 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-t2vw2" podUID="8ef62ffe-f518-4f9e-86fb-8be7f18ef773" containerName="registry-server" containerID="cri-o://3d2d5d7b87c0a19fe20b26558588ec17d316aa916aa594995a31509e4d76c00b" gracePeriod=2 Feb 02 15:39:39 crc kubenswrapper[4733]: I0202 15:39:39.867756 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t2vw2" Feb 02 15:39:40 crc kubenswrapper[4733]: I0202 15:39:40.038956 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ef62ffe-f518-4f9e-86fb-8be7f18ef773-utilities\") pod \"8ef62ffe-f518-4f9e-86fb-8be7f18ef773\" (UID: \"8ef62ffe-f518-4f9e-86fb-8be7f18ef773\") " Feb 02 15:39:40 crc kubenswrapper[4733]: I0202 15:39:40.039018 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ef62ffe-f518-4f9e-86fb-8be7f18ef773-catalog-content\") pod \"8ef62ffe-f518-4f9e-86fb-8be7f18ef773\" (UID: \"8ef62ffe-f518-4f9e-86fb-8be7f18ef773\") " Feb 02 15:39:40 crc kubenswrapper[4733]: I0202 15:39:40.039059 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cx9hp\" (UniqueName: \"kubernetes.io/projected/8ef62ffe-f518-4f9e-86fb-8be7f18ef773-kube-api-access-cx9hp\") pod \"8ef62ffe-f518-4f9e-86fb-8be7f18ef773\" (UID: \"8ef62ffe-f518-4f9e-86fb-8be7f18ef773\") " Feb 02 15:39:40 crc kubenswrapper[4733]: I0202 15:39:40.039968 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ef62ffe-f518-4f9e-86fb-8be7f18ef773-utilities" (OuterVolumeSpecName: "utilities") pod "8ef62ffe-f518-4f9e-86fb-8be7f18ef773" (UID: "8ef62ffe-f518-4f9e-86fb-8be7f18ef773"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:39:40 crc kubenswrapper[4733]: I0202 15:39:40.049489 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ef62ffe-f518-4f9e-86fb-8be7f18ef773-kube-api-access-cx9hp" (OuterVolumeSpecName: "kube-api-access-cx9hp") pod "8ef62ffe-f518-4f9e-86fb-8be7f18ef773" (UID: "8ef62ffe-f518-4f9e-86fb-8be7f18ef773"). InnerVolumeSpecName "kube-api-access-cx9hp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:39:40 crc kubenswrapper[4733]: I0202 15:39:40.098358 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ef62ffe-f518-4f9e-86fb-8be7f18ef773-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8ef62ffe-f518-4f9e-86fb-8be7f18ef773" (UID: "8ef62ffe-f518-4f9e-86fb-8be7f18ef773"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:39:40 crc kubenswrapper[4733]: I0202 15:39:40.141021 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ef62ffe-f518-4f9e-86fb-8be7f18ef773-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 15:39:40 crc kubenswrapper[4733]: I0202 15:39:40.141070 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ef62ffe-f518-4f9e-86fb-8be7f18ef773-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 15:39:40 crc kubenswrapper[4733]: I0202 15:39:40.141089 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cx9hp\" (UniqueName: \"kubernetes.io/projected/8ef62ffe-f518-4f9e-86fb-8be7f18ef773-kube-api-access-cx9hp\") on node \"crc\" DevicePath \"\"" Feb 02 15:39:40 crc kubenswrapper[4733]: I0202 15:39:40.412849 4733 generic.go:334] "Generic (PLEG): container finished" podID="8ef62ffe-f518-4f9e-86fb-8be7f18ef773" containerID="3d2d5d7b87c0a19fe20b26558588ec17d316aa916aa594995a31509e4d76c00b" exitCode=0 Feb 02 15:39:40 crc kubenswrapper[4733]: I0202 15:39:40.412921 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t2vw2" event={"ID":"8ef62ffe-f518-4f9e-86fb-8be7f18ef773","Type":"ContainerDied","Data":"3d2d5d7b87c0a19fe20b26558588ec17d316aa916aa594995a31509e4d76c00b"} Feb 02 15:39:40 crc kubenswrapper[4733]: I0202 15:39:40.412956 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t2vw2" Feb 02 15:39:40 crc kubenswrapper[4733]: I0202 15:39:40.412981 4733 scope.go:117] "RemoveContainer" containerID="3d2d5d7b87c0a19fe20b26558588ec17d316aa916aa594995a31509e4d76c00b" Feb 02 15:39:40 crc kubenswrapper[4733]: I0202 15:39:40.412963 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t2vw2" event={"ID":"8ef62ffe-f518-4f9e-86fb-8be7f18ef773","Type":"ContainerDied","Data":"f3706c857d148eb9d0f3c15115295714fd60d41c38777689a81b0f7e4b80e0af"} Feb 02 15:39:40 crc kubenswrapper[4733]: I0202 15:39:40.446629 4733 scope.go:117] "RemoveContainer" containerID="a7766ebd44e9718fcce4b6ba02ff3771ae6a21da6f35443e6c802a2360018967" Feb 02 15:39:40 crc kubenswrapper[4733]: I0202 15:39:40.464919 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-t2vw2"] Feb 02 15:39:40 crc kubenswrapper[4733]: I0202 15:39:40.471104 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-t2vw2"] Feb 02 15:39:40 crc kubenswrapper[4733]: I0202 15:39:40.486141 4733 scope.go:117] "RemoveContainer" containerID="df3f4e67398d7e7311ebdc12020fee8953b18a916f1565c8f7f84ec249a5a11b" Feb 02 15:39:40 crc kubenswrapper[4733]: I0202 15:39:40.510021 4733 scope.go:117] "RemoveContainer" containerID="3d2d5d7b87c0a19fe20b26558588ec17d316aa916aa594995a31509e4d76c00b" Feb 02 15:39:40 crc kubenswrapper[4733]: E0202 15:39:40.511580 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d2d5d7b87c0a19fe20b26558588ec17d316aa916aa594995a31509e4d76c00b\": container with ID starting with 3d2d5d7b87c0a19fe20b26558588ec17d316aa916aa594995a31509e4d76c00b not found: ID does not exist" containerID="3d2d5d7b87c0a19fe20b26558588ec17d316aa916aa594995a31509e4d76c00b" Feb 02 15:39:40 crc kubenswrapper[4733]: I0202 15:39:40.511690 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d2d5d7b87c0a19fe20b26558588ec17d316aa916aa594995a31509e4d76c00b"} err="failed to get container status \"3d2d5d7b87c0a19fe20b26558588ec17d316aa916aa594995a31509e4d76c00b\": rpc error: code = NotFound desc = could not find container \"3d2d5d7b87c0a19fe20b26558588ec17d316aa916aa594995a31509e4d76c00b\": container with ID starting with 3d2d5d7b87c0a19fe20b26558588ec17d316aa916aa594995a31509e4d76c00b not found: ID does not exist" Feb 02 15:39:40 crc kubenswrapper[4733]: I0202 15:39:40.511773 4733 scope.go:117] "RemoveContainer" containerID="a7766ebd44e9718fcce4b6ba02ff3771ae6a21da6f35443e6c802a2360018967" Feb 02 15:39:40 crc kubenswrapper[4733]: E0202 15:39:40.512336 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7766ebd44e9718fcce4b6ba02ff3771ae6a21da6f35443e6c802a2360018967\": container with ID starting with a7766ebd44e9718fcce4b6ba02ff3771ae6a21da6f35443e6c802a2360018967 not found: ID does not exist" containerID="a7766ebd44e9718fcce4b6ba02ff3771ae6a21da6f35443e6c802a2360018967" Feb 02 15:39:40 crc kubenswrapper[4733]: I0202 15:39:40.512408 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7766ebd44e9718fcce4b6ba02ff3771ae6a21da6f35443e6c802a2360018967"} err="failed to get container status \"a7766ebd44e9718fcce4b6ba02ff3771ae6a21da6f35443e6c802a2360018967\": rpc error: code = NotFound desc = could not find container \"a7766ebd44e9718fcce4b6ba02ff3771ae6a21da6f35443e6c802a2360018967\": container with ID starting with a7766ebd44e9718fcce4b6ba02ff3771ae6a21da6f35443e6c802a2360018967 not found: ID does not exist" Feb 02 15:39:40 crc kubenswrapper[4733]: I0202 15:39:40.512451 4733 scope.go:117] "RemoveContainer" containerID="df3f4e67398d7e7311ebdc12020fee8953b18a916f1565c8f7f84ec249a5a11b" Feb 02 15:39:40 crc kubenswrapper[4733]: E0202 15:39:40.512982 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df3f4e67398d7e7311ebdc12020fee8953b18a916f1565c8f7f84ec249a5a11b\": container with ID starting with df3f4e67398d7e7311ebdc12020fee8953b18a916f1565c8f7f84ec249a5a11b not found: ID does not exist" containerID="df3f4e67398d7e7311ebdc12020fee8953b18a916f1565c8f7f84ec249a5a11b" Feb 02 15:39:40 crc kubenswrapper[4733]: I0202 15:39:40.513132 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df3f4e67398d7e7311ebdc12020fee8953b18a916f1565c8f7f84ec249a5a11b"} err="failed to get container status \"df3f4e67398d7e7311ebdc12020fee8953b18a916f1565c8f7f84ec249a5a11b\": rpc error: code = NotFound desc = could not find container \"df3f4e67398d7e7311ebdc12020fee8953b18a916f1565c8f7f84ec249a5a11b\": container with ID starting with df3f4e67398d7e7311ebdc12020fee8953b18a916f1565c8f7f84ec249a5a11b not found: ID does not exist" Feb 02 15:39:41 crc kubenswrapper[4733]: I0202 15:39:41.271987 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ef62ffe-f518-4f9e-86fb-8be7f18ef773" path="/var/lib/kubelet/pods/8ef62ffe-f518-4f9e-86fb-8be7f18ef773/volumes" Feb 02 15:39:44 crc kubenswrapper[4733]: I0202 15:39:44.275128 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-4hscm/must-gather-thtgn"] Feb 02 15:39:44 crc kubenswrapper[4733]: E0202 15:39:44.275861 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ef62ffe-f518-4f9e-86fb-8be7f18ef773" containerName="extract-content" Feb 02 15:39:44 crc kubenswrapper[4733]: I0202 15:39:44.275873 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ef62ffe-f518-4f9e-86fb-8be7f18ef773" containerName="extract-content" Feb 02 15:39:44 crc kubenswrapper[4733]: E0202 15:39:44.275889 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ef62ffe-f518-4f9e-86fb-8be7f18ef773" containerName="registry-server" Feb 02 15:39:44 crc kubenswrapper[4733]: I0202 15:39:44.275895 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ef62ffe-f518-4f9e-86fb-8be7f18ef773" containerName="registry-server" Feb 02 15:39:44 crc kubenswrapper[4733]: E0202 15:39:44.275908 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ef62ffe-f518-4f9e-86fb-8be7f18ef773" containerName="extract-utilities" Feb 02 15:39:44 crc kubenswrapper[4733]: I0202 15:39:44.275915 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ef62ffe-f518-4f9e-86fb-8be7f18ef773" containerName="extract-utilities" Feb 02 15:39:44 crc kubenswrapper[4733]: I0202 15:39:44.276028 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ef62ffe-f518-4f9e-86fb-8be7f18ef773" containerName="registry-server" Feb 02 15:39:44 crc kubenswrapper[4733]: I0202 15:39:44.276595 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4hscm/must-gather-thtgn" Feb 02 15:39:44 crc kubenswrapper[4733]: I0202 15:39:44.278569 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-4hscm"/"default-dockercfg-qx6c4" Feb 02 15:39:44 crc kubenswrapper[4733]: I0202 15:39:44.280356 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-4hscm"/"openshift-service-ca.crt" Feb 02 15:39:44 crc kubenswrapper[4733]: I0202 15:39:44.280667 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-4hscm"/"kube-root-ca.crt" Feb 02 15:39:44 crc kubenswrapper[4733]: I0202 15:39:44.284961 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-4hscm/must-gather-thtgn"] Feb 02 15:39:44 crc kubenswrapper[4733]: I0202 15:39:44.305030 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdx6n\" (UniqueName: \"kubernetes.io/projected/5893b4e6-a11b-4e61-ba80-2ab3963863db-kube-api-access-vdx6n\") pod \"must-gather-thtgn\" (UID: \"5893b4e6-a11b-4e61-ba80-2ab3963863db\") " pod="openshift-must-gather-4hscm/must-gather-thtgn" Feb 02 15:39:44 crc kubenswrapper[4733]: I0202 15:39:44.305109 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5893b4e6-a11b-4e61-ba80-2ab3963863db-must-gather-output\") pod \"must-gather-thtgn\" (UID: \"5893b4e6-a11b-4e61-ba80-2ab3963863db\") " pod="openshift-must-gather-4hscm/must-gather-thtgn" Feb 02 15:39:44 crc kubenswrapper[4733]: I0202 15:39:44.406045 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdx6n\" (UniqueName: \"kubernetes.io/projected/5893b4e6-a11b-4e61-ba80-2ab3963863db-kube-api-access-vdx6n\") pod \"must-gather-thtgn\" (UID: \"5893b4e6-a11b-4e61-ba80-2ab3963863db\") " pod="openshift-must-gather-4hscm/must-gather-thtgn" Feb 02 15:39:44 crc kubenswrapper[4733]: I0202 15:39:44.406100 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5893b4e6-a11b-4e61-ba80-2ab3963863db-must-gather-output\") pod \"must-gather-thtgn\" (UID: \"5893b4e6-a11b-4e61-ba80-2ab3963863db\") " pod="openshift-must-gather-4hscm/must-gather-thtgn" Feb 02 15:39:44 crc kubenswrapper[4733]: I0202 15:39:44.406681 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5893b4e6-a11b-4e61-ba80-2ab3963863db-must-gather-output\") pod \"must-gather-thtgn\" (UID: \"5893b4e6-a11b-4e61-ba80-2ab3963863db\") " pod="openshift-must-gather-4hscm/must-gather-thtgn" Feb 02 15:39:44 crc kubenswrapper[4733]: I0202 15:39:44.423381 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdx6n\" (UniqueName: \"kubernetes.io/projected/5893b4e6-a11b-4e61-ba80-2ab3963863db-kube-api-access-vdx6n\") pod \"must-gather-thtgn\" (UID: \"5893b4e6-a11b-4e61-ba80-2ab3963863db\") " pod="openshift-must-gather-4hscm/must-gather-thtgn" Feb 02 15:39:44 crc kubenswrapper[4733]: I0202 15:39:44.589807 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4hscm/must-gather-thtgn" Feb 02 15:39:44 crc kubenswrapper[4733]: I0202 15:39:44.799104 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-4hscm/must-gather-thtgn"] Feb 02 15:39:44 crc kubenswrapper[4733]: W0202 15:39:44.805049 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5893b4e6_a11b_4e61_ba80_2ab3963863db.slice/crio-d88943d3e009ab86afc953adb2a2b8e5586dc92b1185a98576b5f93ff101c2f4 WatchSource:0}: Error finding container d88943d3e009ab86afc953adb2a2b8e5586dc92b1185a98576b5f93ff101c2f4: Status 404 returned error can't find the container with id d88943d3e009ab86afc953adb2a2b8e5586dc92b1185a98576b5f93ff101c2f4 Feb 02 15:39:45 crc kubenswrapper[4733]: I0202 15:39:45.451694 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4hscm/must-gather-thtgn" event={"ID":"5893b4e6-a11b-4e61-ba80-2ab3963863db","Type":"ContainerStarted","Data":"d88943d3e009ab86afc953adb2a2b8e5586dc92b1185a98576b5f93ff101c2f4"} Feb 02 15:39:49 crc kubenswrapper[4733]: I0202 15:39:49.497632 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4hscm/must-gather-thtgn" event={"ID":"5893b4e6-a11b-4e61-ba80-2ab3963863db","Type":"ContainerStarted","Data":"f8165844c6b7202570e0d95ddce2266894640e7e2835b3d6d5a52e66c1bfa122"} Feb 02 15:39:50 crc kubenswrapper[4733]: I0202 15:39:50.505881 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4hscm/must-gather-thtgn" event={"ID":"5893b4e6-a11b-4e61-ba80-2ab3963863db","Type":"ContainerStarted","Data":"f7460dc02ba64f0b5cbb3f9e6614ae86acf755fea2961e1167fb04f7fd05bf32"} Feb 02 15:39:50 crc kubenswrapper[4733]: I0202 15:39:50.523575 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-4hscm/must-gather-thtgn" podStartSLOduration=2.210811532 podStartE2EDuration="6.523556195s" podCreationTimestamp="2026-02-02 15:39:44 +0000 UTC" firstStartedPulling="2026-02-02 15:39:44.809646211 +0000 UTC m=+1468.261107569" lastFinishedPulling="2026-02-02 15:39:49.122390864 +0000 UTC m=+1472.573852232" observedRunningTime="2026-02-02 15:39:50.523444882 +0000 UTC m=+1473.974906280" watchObservedRunningTime="2026-02-02 15:39:50.523556195 +0000 UTC m=+1473.975017563" Feb 02 15:40:04 crc kubenswrapper[4733]: I0202 15:40:04.987716 4733 patch_prober.go:28] interesting pod/machine-config-daemon-zgsv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 15:40:04 crc kubenswrapper[4733]: I0202 15:40:04.988413 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 15:40:04 crc kubenswrapper[4733]: I0202 15:40:04.988479 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" Feb 02 15:40:04 crc kubenswrapper[4733]: I0202 15:40:04.989383 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9ef5e0247a233902273cc6edac5d4e35e73d0f3d5ded3fb8bb539db06e43f773"} pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 15:40:04 crc kubenswrapper[4733]: I0202 15:40:04.989503 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" containerName="machine-config-daemon" containerID="cri-o://9ef5e0247a233902273cc6edac5d4e35e73d0f3d5ded3fb8bb539db06e43f773" gracePeriod=600 Feb 02 15:40:05 crc kubenswrapper[4733]: I0202 15:40:05.626219 4733 generic.go:334] "Generic (PLEG): container finished" podID="83d0e810-ea5c-47aa-947b-7865b045bf94" containerID="9ef5e0247a233902273cc6edac5d4e35e73d0f3d5ded3fb8bb539db06e43f773" exitCode=0 Feb 02 15:40:05 crc kubenswrapper[4733]: I0202 15:40:05.626295 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" event={"ID":"83d0e810-ea5c-47aa-947b-7865b045bf94","Type":"ContainerDied","Data":"9ef5e0247a233902273cc6edac5d4e35e73d0f3d5ded3fb8bb539db06e43f773"} Feb 02 15:40:05 crc kubenswrapper[4733]: I0202 15:40:05.626593 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" event={"ID":"83d0e810-ea5c-47aa-947b-7865b045bf94","Type":"ContainerStarted","Data":"8837c504005a41d8189bc30082d6c198635f60c659df3e0066907bf203ae6b97"} Feb 02 15:40:05 crc kubenswrapper[4733]: I0202 15:40:05.626625 4733 scope.go:117] "RemoveContainer" containerID="26cec88244c10ffe50680f654ac54834881498bea583f15ab3590dda3791c1bf" Feb 02 15:40:19 crc kubenswrapper[4733]: I0202 15:40:19.949619 4733 scope.go:117] "RemoveContainer" containerID="bdbb55dce4f250f0293be7e174c78454be12f47e5dd826641c4dd8a005f5bf3f" Feb 02 15:40:19 crc kubenswrapper[4733]: I0202 15:40:19.990494 4733 scope.go:117] "RemoveContainer" containerID="6b9792c10e3b3d8117ac5087d0b243117b2689263cdeec866a2c3fe00d543847" Feb 02 15:40:20 crc kubenswrapper[4733]: I0202 15:40:20.017453 4733 scope.go:117] "RemoveContainer" containerID="2bacaed5ef8ccec41d0fbf7135bc788ffa9e877f1c64af132fe7982b651e1f67" Feb 02 15:40:20 crc kubenswrapper[4733]: I0202 15:40:20.054009 4733 scope.go:117] "RemoveContainer" containerID="d4696ce0ed2dab71f0282caae5dd7bafb4b5e9a7b706a558a2cca872fa7edde3" Feb 02 15:40:20 crc kubenswrapper[4733]: I0202 15:40:20.088295 4733 scope.go:117] "RemoveContainer" containerID="a95fee8d9d9b365d3b0c6597d3f5027e679d594c5a1448b4a40af3af0022b2f2" Feb 02 15:40:20 crc kubenswrapper[4733]: I0202 15:40:20.116588 4733 scope.go:117] "RemoveContainer" containerID="c8112f53db575592095d0a3ac1b3f517d2cd02c636461f258aad664c4709316f" Feb 02 15:40:20 crc kubenswrapper[4733]: I0202 15:40:20.140854 4733 scope.go:117] "RemoveContainer" containerID="77f034135db5431b38fb66fe9ca3c5c0eb6838124fe6c9d890a8238d3715408a" Feb 02 15:40:20 crc kubenswrapper[4733]: I0202 15:40:20.157885 4733 scope.go:117] "RemoveContainer" containerID="ab19f2f2d434db0186b91ba8763d53b716f7e1126e222060a31f15c1c270df31" Feb 02 15:40:20 crc kubenswrapper[4733]: I0202 15:40:20.179303 4733 scope.go:117] "RemoveContainer" containerID="1d6342cba8f24fd454efd3c05383711f7bc7a5f9589dc436535e902f27915680" Feb 02 15:40:39 crc kubenswrapper[4733]: I0202 15:40:39.574238 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-dkktg_bd697754-4524-44ed-be19-18747c0626f5/control-plane-machine-set-operator/0.log" Feb 02 15:40:39 crc kubenswrapper[4733]: I0202 15:40:39.746076 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-74k4s_a3e78085-1f5d-4f0c-84bb-62b787b1fbdd/kube-rbac-proxy/0.log" Feb 02 15:40:39 crc kubenswrapper[4733]: I0202 15:40:39.764926 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-74k4s_a3e78085-1f5d-4f0c-84bb-62b787b1fbdd/machine-api-operator/0.log" Feb 02 15:41:08 crc kubenswrapper[4733]: I0202 15:41:08.861105 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-h9fwf_222b6fa1-d59f-4779-8fdf-72c3fd3ea2cc/controller/0.log" Feb 02 15:41:08 crc kubenswrapper[4733]: I0202 15:41:08.884709 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-h9fwf_222b6fa1-d59f-4779-8fdf-72c3fd3ea2cc/kube-rbac-proxy/0.log" Feb 02 15:41:09 crc kubenswrapper[4733]: I0202 15:41:09.008443 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l6zg6_915914ae-77a5-4bf7-9d71-6207b62d2f50/cp-frr-files/0.log" Feb 02 15:41:09 crc kubenswrapper[4733]: I0202 15:41:09.161935 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l6zg6_915914ae-77a5-4bf7-9d71-6207b62d2f50/cp-reloader/0.log" Feb 02 15:41:09 crc kubenswrapper[4733]: I0202 15:41:09.194037 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l6zg6_915914ae-77a5-4bf7-9d71-6207b62d2f50/cp-metrics/0.log" Feb 02 15:41:09 crc kubenswrapper[4733]: I0202 15:41:09.212398 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l6zg6_915914ae-77a5-4bf7-9d71-6207b62d2f50/cp-frr-files/0.log" Feb 02 15:41:09 crc kubenswrapper[4733]: I0202 15:41:09.258108 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l6zg6_915914ae-77a5-4bf7-9d71-6207b62d2f50/cp-reloader/0.log" Feb 02 15:41:09 crc kubenswrapper[4733]: I0202 15:41:09.419018 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l6zg6_915914ae-77a5-4bf7-9d71-6207b62d2f50/cp-metrics/0.log" Feb 02 15:41:09 crc kubenswrapper[4733]: I0202 15:41:09.419046 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l6zg6_915914ae-77a5-4bf7-9d71-6207b62d2f50/cp-frr-files/0.log" Feb 02 15:41:09 crc kubenswrapper[4733]: I0202 15:41:09.432853 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l6zg6_915914ae-77a5-4bf7-9d71-6207b62d2f50/cp-metrics/0.log" Feb 02 15:41:09 crc kubenswrapper[4733]: I0202 15:41:09.460423 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l6zg6_915914ae-77a5-4bf7-9d71-6207b62d2f50/cp-reloader/0.log" Feb 02 15:41:09 crc kubenswrapper[4733]: I0202 15:41:09.598082 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l6zg6_915914ae-77a5-4bf7-9d71-6207b62d2f50/cp-frr-files/0.log" Feb 02 15:41:09 crc kubenswrapper[4733]: I0202 15:41:09.617245 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l6zg6_915914ae-77a5-4bf7-9d71-6207b62d2f50/cp-reloader/0.log" Feb 02 15:41:09 crc kubenswrapper[4733]: I0202 15:41:09.622293 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l6zg6_915914ae-77a5-4bf7-9d71-6207b62d2f50/cp-metrics/0.log" Feb 02 15:41:09 crc kubenswrapper[4733]: I0202 15:41:09.653753 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l6zg6_915914ae-77a5-4bf7-9d71-6207b62d2f50/controller/0.log" Feb 02 15:41:09 crc kubenswrapper[4733]: I0202 15:41:09.800131 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l6zg6_915914ae-77a5-4bf7-9d71-6207b62d2f50/frr-metrics/0.log" Feb 02 15:41:09 crc kubenswrapper[4733]: I0202 15:41:09.800670 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l6zg6_915914ae-77a5-4bf7-9d71-6207b62d2f50/kube-rbac-proxy/0.log" Feb 02 15:41:09 crc kubenswrapper[4733]: I0202 15:41:09.832528 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l6zg6_915914ae-77a5-4bf7-9d71-6207b62d2f50/kube-rbac-proxy-frr/0.log" Feb 02 15:41:09 crc kubenswrapper[4733]: I0202 15:41:09.983340 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l6zg6_915914ae-77a5-4bf7-9d71-6207b62d2f50/reloader/0.log" Feb 02 15:41:10 crc kubenswrapper[4733]: I0202 15:41:10.035053 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7df86c4f6c-5l6sj_70e35df4-ce17-4181-a64c-d7a70160d0e7/frr-k8s-webhook-server/0.log" Feb 02 15:41:10 crc kubenswrapper[4733]: I0202 15:41:10.195394 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-54b6695bd-mwgx2_2e7a0012-8704-4613-9643-ccb780c9d76d/manager/0.log" Feb 02 15:41:10 crc kubenswrapper[4733]: I0202 15:41:10.224000 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l6zg6_915914ae-77a5-4bf7-9d71-6207b62d2f50/frr/0.log" Feb 02 15:41:10 crc kubenswrapper[4733]: I0202 15:41:10.339406 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-66bb599b85-wt9zk_ffb4086a-a662-4380-8dd0-c218d36e7c31/webhook-server/0.log" Feb 02 15:41:10 crc kubenswrapper[4733]: I0202 15:41:10.393863 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-7pnkj_1c5ad065-b35b-4784-89b6-72e8dd0a1df5/kube-rbac-proxy/0.log" Feb 02 15:41:10 crc kubenswrapper[4733]: I0202 15:41:10.529299 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-7pnkj_1c5ad065-b35b-4784-89b6-72e8dd0a1df5/speaker/0.log" Feb 02 15:41:20 crc kubenswrapper[4733]: I0202 15:41:20.283801 4733 scope.go:117] "RemoveContainer" containerID="ad074bd93f775844cad04e7d9ac407c52dca41251c0106fba43a6c42de4a1e7e" Feb 02 15:41:20 crc kubenswrapper[4733]: I0202 15:41:20.320524 4733 scope.go:117] "RemoveContainer" containerID="09989f03d83fa6e83049f56a1c6e1ae9819f524cb9c149681d71df8727e1096e" Feb 02 15:41:20 crc kubenswrapper[4733]: I0202 15:41:20.345777 4733 scope.go:117] "RemoveContainer" containerID="16745d57b94a753d9e93ffa48cc9c02bda6166e8d95e1a988a017658b15ed912" Feb 02 15:41:36 crc kubenswrapper[4733]: I0202 15:41:36.202637 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg_e8c251f4-787d-49f1-8ec3-32f6c5dff2bb/util/0.log" Feb 02 15:41:36 crc kubenswrapper[4733]: I0202 15:41:36.286875 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg_e8c251f4-787d-49f1-8ec3-32f6c5dff2bb/util/0.log" Feb 02 15:41:36 crc kubenswrapper[4733]: I0202 15:41:36.348140 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg_e8c251f4-787d-49f1-8ec3-32f6c5dff2bb/pull/0.log" Feb 02 15:41:36 crc kubenswrapper[4733]: I0202 15:41:36.354701 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg_e8c251f4-787d-49f1-8ec3-32f6c5dff2bb/pull/0.log" Feb 02 15:41:36 crc kubenswrapper[4733]: I0202 15:41:36.549403 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg_e8c251f4-787d-49f1-8ec3-32f6c5dff2bb/extract/0.log" Feb 02 15:41:36 crc kubenswrapper[4733]: I0202 15:41:36.572259 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg_e8c251f4-787d-49f1-8ec3-32f6c5dff2bb/util/0.log" Feb 02 15:41:36 crc kubenswrapper[4733]: I0202 15:41:36.587549 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg_e8c251f4-787d-49f1-8ec3-32f6c5dff2bb/pull/0.log" Feb 02 15:41:36 crc kubenswrapper[4733]: I0202 15:41:36.697151 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mbh5r_b8fd49b0-ac5d-4b93-8c03-7ac35c150aa2/extract-utilities/0.log" Feb 02 15:41:36 crc kubenswrapper[4733]: I0202 15:41:36.900355 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mbh5r_b8fd49b0-ac5d-4b93-8c03-7ac35c150aa2/extract-content/0.log" Feb 02 15:41:36 crc kubenswrapper[4733]: I0202 15:41:36.907652 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mbh5r_b8fd49b0-ac5d-4b93-8c03-7ac35c150aa2/extract-utilities/0.log" Feb 02 15:41:36 crc kubenswrapper[4733]: I0202 15:41:36.907695 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mbh5r_b8fd49b0-ac5d-4b93-8c03-7ac35c150aa2/extract-content/0.log" Feb 02 15:41:37 crc kubenswrapper[4733]: I0202 15:41:37.094321 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mbh5r_b8fd49b0-ac5d-4b93-8c03-7ac35c150aa2/extract-content/0.log" Feb 02 15:41:37 crc kubenswrapper[4733]: I0202 15:41:37.096000 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mbh5r_b8fd49b0-ac5d-4b93-8c03-7ac35c150aa2/extract-utilities/0.log" Feb 02 15:41:37 crc kubenswrapper[4733]: I0202 15:41:37.291089 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qvwpw_54e05a7f-a312-4455-83ec-be97bf446b68/extract-utilities/0.log" Feb 02 15:41:37 crc kubenswrapper[4733]: I0202 15:41:37.419979 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mbh5r_b8fd49b0-ac5d-4b93-8c03-7ac35c150aa2/registry-server/0.log" Feb 02 15:41:37 crc kubenswrapper[4733]: I0202 15:41:37.451500 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qvwpw_54e05a7f-a312-4455-83ec-be97bf446b68/extract-content/0.log" Feb 02 15:41:37 crc kubenswrapper[4733]: I0202 15:41:37.455425 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qvwpw_54e05a7f-a312-4455-83ec-be97bf446b68/extract-utilities/0.log" Feb 02 15:41:37 crc kubenswrapper[4733]: I0202 15:41:37.490419 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qvwpw_54e05a7f-a312-4455-83ec-be97bf446b68/extract-content/0.log" Feb 02 15:41:37 crc kubenswrapper[4733]: I0202 15:41:37.685968 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qvwpw_54e05a7f-a312-4455-83ec-be97bf446b68/extract-content/0.log" Feb 02 15:41:37 crc kubenswrapper[4733]: I0202 15:41:37.688742 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qvwpw_54e05a7f-a312-4455-83ec-be97bf446b68/extract-utilities/0.log" Feb 02 15:41:37 crc kubenswrapper[4733]: I0202 15:41:37.846436 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-ps6dr_babbc7b3-6e09-43a7-9cbc-70ba7076ab9f/marketplace-operator/0.log" Feb 02 15:41:37 crc kubenswrapper[4733]: I0202 15:41:37.902485 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-smvdc_0367d1cb-4122-4c7e-b203-d5c4216c6ef5/extract-utilities/0.log" Feb 02 15:41:37 crc kubenswrapper[4733]: I0202 15:41:37.959828 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qvwpw_54e05a7f-a312-4455-83ec-be97bf446b68/registry-server/0.log" Feb 02 15:41:38 crc kubenswrapper[4733]: I0202 15:41:38.041925 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-smvdc_0367d1cb-4122-4c7e-b203-d5c4216c6ef5/extract-utilities/0.log" Feb 02 15:41:38 crc kubenswrapper[4733]: I0202 15:41:38.074586 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-smvdc_0367d1cb-4122-4c7e-b203-d5c4216c6ef5/extract-content/0.log" Feb 02 15:41:38 crc kubenswrapper[4733]: I0202 15:41:38.080515 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-smvdc_0367d1cb-4122-4c7e-b203-d5c4216c6ef5/extract-content/0.log" Feb 02 15:41:38 crc kubenswrapper[4733]: I0202 15:41:38.261035 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-smvdc_0367d1cb-4122-4c7e-b203-d5c4216c6ef5/extract-content/0.log" Feb 02 15:41:38 crc kubenswrapper[4733]: I0202 15:41:38.269447 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-smvdc_0367d1cb-4122-4c7e-b203-d5c4216c6ef5/extract-utilities/0.log" Feb 02 15:41:38 crc kubenswrapper[4733]: I0202 15:41:38.312326 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-smvdc_0367d1cb-4122-4c7e-b203-d5c4216c6ef5/registry-server/0.log" Feb 02 15:41:38 crc kubenswrapper[4733]: I0202 15:41:38.448717 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-kqbhk_bcccf11d-2567-42d7-81e2-5684b4b43823/extract-utilities/0.log" Feb 02 15:41:38 crc kubenswrapper[4733]: I0202 15:41:38.572102 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-kqbhk_bcccf11d-2567-42d7-81e2-5684b4b43823/extract-utilities/0.log" Feb 02 15:41:38 crc kubenswrapper[4733]: I0202 15:41:38.577764 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-kqbhk_bcccf11d-2567-42d7-81e2-5684b4b43823/extract-content/0.log" Feb 02 15:41:38 crc kubenswrapper[4733]: I0202 15:41:38.619253 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-kqbhk_bcccf11d-2567-42d7-81e2-5684b4b43823/extract-content/0.log" Feb 02 15:41:38 crc kubenswrapper[4733]: I0202 15:41:38.757574 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-kqbhk_bcccf11d-2567-42d7-81e2-5684b4b43823/extract-utilities/0.log" Feb 02 15:41:38 crc kubenswrapper[4733]: I0202 15:41:38.762215 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-kqbhk_bcccf11d-2567-42d7-81e2-5684b4b43823/extract-content/0.log" Feb 02 15:41:39 crc kubenswrapper[4733]: I0202 15:41:39.059002 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-kqbhk_bcccf11d-2567-42d7-81e2-5684b4b43823/registry-server/0.log" Feb 02 15:42:34 crc kubenswrapper[4733]: I0202 15:42:34.988122 4733 patch_prober.go:28] interesting pod/machine-config-daemon-zgsv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 15:42:34 crc kubenswrapper[4733]: I0202 15:42:34.988739 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 15:42:48 crc kubenswrapper[4733]: I0202 15:42:48.751373 4733 generic.go:334] "Generic (PLEG): container finished" podID="5893b4e6-a11b-4e61-ba80-2ab3963863db" containerID="f8165844c6b7202570e0d95ddce2266894640e7e2835b3d6d5a52e66c1bfa122" exitCode=0 Feb 02 15:42:48 crc kubenswrapper[4733]: I0202 15:42:48.751518 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4hscm/must-gather-thtgn" event={"ID":"5893b4e6-a11b-4e61-ba80-2ab3963863db","Type":"ContainerDied","Data":"f8165844c6b7202570e0d95ddce2266894640e7e2835b3d6d5a52e66c1bfa122"} Feb 02 15:42:48 crc kubenswrapper[4733]: I0202 15:42:48.752791 4733 scope.go:117] "RemoveContainer" containerID="f8165844c6b7202570e0d95ddce2266894640e7e2835b3d6d5a52e66c1bfa122" Feb 02 15:42:49 crc kubenswrapper[4733]: I0202 15:42:49.151042 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-4hscm_must-gather-thtgn_5893b4e6-a11b-4e61-ba80-2ab3963863db/gather/0.log" Feb 02 15:42:56 crc kubenswrapper[4733]: I0202 15:42:56.021560 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-4hscm/must-gather-thtgn"] Feb 02 15:42:56 crc kubenswrapper[4733]: I0202 15:42:56.022336 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-4hscm/must-gather-thtgn" podUID="5893b4e6-a11b-4e61-ba80-2ab3963863db" containerName="copy" containerID="cri-o://f7460dc02ba64f0b5cbb3f9e6614ae86acf755fea2961e1167fb04f7fd05bf32" gracePeriod=2 Feb 02 15:42:56 crc kubenswrapper[4733]: I0202 15:42:56.028897 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-4hscm/must-gather-thtgn"] Feb 02 15:42:56 crc kubenswrapper[4733]: I0202 15:42:56.412184 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-4hscm_must-gather-thtgn_5893b4e6-a11b-4e61-ba80-2ab3963863db/copy/0.log" Feb 02 15:42:56 crc kubenswrapper[4733]: I0202 15:42:56.413061 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4hscm/must-gather-thtgn" Feb 02 15:42:56 crc kubenswrapper[4733]: I0202 15:42:56.482722 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vdx6n\" (UniqueName: \"kubernetes.io/projected/5893b4e6-a11b-4e61-ba80-2ab3963863db-kube-api-access-vdx6n\") pod \"5893b4e6-a11b-4e61-ba80-2ab3963863db\" (UID: \"5893b4e6-a11b-4e61-ba80-2ab3963863db\") " Feb 02 15:42:56 crc kubenswrapper[4733]: I0202 15:42:56.482795 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5893b4e6-a11b-4e61-ba80-2ab3963863db-must-gather-output\") pod \"5893b4e6-a11b-4e61-ba80-2ab3963863db\" (UID: \"5893b4e6-a11b-4e61-ba80-2ab3963863db\") " Feb 02 15:42:56 crc kubenswrapper[4733]: I0202 15:42:56.488774 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5893b4e6-a11b-4e61-ba80-2ab3963863db-kube-api-access-vdx6n" (OuterVolumeSpecName: "kube-api-access-vdx6n") pod "5893b4e6-a11b-4e61-ba80-2ab3963863db" (UID: "5893b4e6-a11b-4e61-ba80-2ab3963863db"). InnerVolumeSpecName "kube-api-access-vdx6n". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:42:56 crc kubenswrapper[4733]: I0202 15:42:56.545494 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5893b4e6-a11b-4e61-ba80-2ab3963863db-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "5893b4e6-a11b-4e61-ba80-2ab3963863db" (UID: "5893b4e6-a11b-4e61-ba80-2ab3963863db"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:42:56 crc kubenswrapper[4733]: I0202 15:42:56.584573 4733 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5893b4e6-a11b-4e61-ba80-2ab3963863db-must-gather-output\") on node \"crc\" DevicePath \"\"" Feb 02 15:42:56 crc kubenswrapper[4733]: I0202 15:42:56.584606 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vdx6n\" (UniqueName: \"kubernetes.io/projected/5893b4e6-a11b-4e61-ba80-2ab3963863db-kube-api-access-vdx6n\") on node \"crc\" DevicePath \"\"" Feb 02 15:42:56 crc kubenswrapper[4733]: I0202 15:42:56.818313 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-4hscm_must-gather-thtgn_5893b4e6-a11b-4e61-ba80-2ab3963863db/copy/0.log" Feb 02 15:42:56 crc kubenswrapper[4733]: I0202 15:42:56.819092 4733 generic.go:334] "Generic (PLEG): container finished" podID="5893b4e6-a11b-4e61-ba80-2ab3963863db" containerID="f7460dc02ba64f0b5cbb3f9e6614ae86acf755fea2961e1167fb04f7fd05bf32" exitCode=143 Feb 02 15:42:56 crc kubenswrapper[4733]: I0202 15:42:56.819232 4733 scope.go:117] "RemoveContainer" containerID="f7460dc02ba64f0b5cbb3f9e6614ae86acf755fea2961e1167fb04f7fd05bf32" Feb 02 15:42:56 crc kubenswrapper[4733]: I0202 15:42:56.819242 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4hscm/must-gather-thtgn" Feb 02 15:42:56 crc kubenswrapper[4733]: I0202 15:42:56.833698 4733 scope.go:117] "RemoveContainer" containerID="f8165844c6b7202570e0d95ddce2266894640e7e2835b3d6d5a52e66c1bfa122" Feb 02 15:42:56 crc kubenswrapper[4733]: I0202 15:42:56.877388 4733 scope.go:117] "RemoveContainer" containerID="f7460dc02ba64f0b5cbb3f9e6614ae86acf755fea2961e1167fb04f7fd05bf32" Feb 02 15:42:56 crc kubenswrapper[4733]: E0202 15:42:56.877944 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7460dc02ba64f0b5cbb3f9e6614ae86acf755fea2961e1167fb04f7fd05bf32\": container with ID starting with f7460dc02ba64f0b5cbb3f9e6614ae86acf755fea2961e1167fb04f7fd05bf32 not found: ID does not exist" containerID="f7460dc02ba64f0b5cbb3f9e6614ae86acf755fea2961e1167fb04f7fd05bf32" Feb 02 15:42:56 crc kubenswrapper[4733]: I0202 15:42:56.878003 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7460dc02ba64f0b5cbb3f9e6614ae86acf755fea2961e1167fb04f7fd05bf32"} err="failed to get container status \"f7460dc02ba64f0b5cbb3f9e6614ae86acf755fea2961e1167fb04f7fd05bf32\": rpc error: code = NotFound desc = could not find container \"f7460dc02ba64f0b5cbb3f9e6614ae86acf755fea2961e1167fb04f7fd05bf32\": container with ID starting with f7460dc02ba64f0b5cbb3f9e6614ae86acf755fea2961e1167fb04f7fd05bf32 not found: ID does not exist" Feb 02 15:42:56 crc kubenswrapper[4733]: I0202 15:42:56.878041 4733 scope.go:117] "RemoveContainer" containerID="f8165844c6b7202570e0d95ddce2266894640e7e2835b3d6d5a52e66c1bfa122" Feb 02 15:42:56 crc kubenswrapper[4733]: E0202 15:42:56.878691 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8165844c6b7202570e0d95ddce2266894640e7e2835b3d6d5a52e66c1bfa122\": container with ID starting with f8165844c6b7202570e0d95ddce2266894640e7e2835b3d6d5a52e66c1bfa122 not found: ID does not exist" containerID="f8165844c6b7202570e0d95ddce2266894640e7e2835b3d6d5a52e66c1bfa122" Feb 02 15:42:56 crc kubenswrapper[4733]: I0202 15:42:56.878754 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8165844c6b7202570e0d95ddce2266894640e7e2835b3d6d5a52e66c1bfa122"} err="failed to get container status \"f8165844c6b7202570e0d95ddce2266894640e7e2835b3d6d5a52e66c1bfa122\": rpc error: code = NotFound desc = could not find container \"f8165844c6b7202570e0d95ddce2266894640e7e2835b3d6d5a52e66c1bfa122\": container with ID starting with f8165844c6b7202570e0d95ddce2266894640e7e2835b3d6d5a52e66c1bfa122 not found: ID does not exist" Feb 02 15:42:57 crc kubenswrapper[4733]: I0202 15:42:57.264026 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5893b4e6-a11b-4e61-ba80-2ab3963863db" path="/var/lib/kubelet/pods/5893b4e6-a11b-4e61-ba80-2ab3963863db/volumes" Feb 02 15:43:04 crc kubenswrapper[4733]: I0202 15:43:04.988041 4733 patch_prober.go:28] interesting pod/machine-config-daemon-zgsv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 15:43:04 crc kubenswrapper[4733]: I0202 15:43:04.988837 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 15:43:20 crc kubenswrapper[4733]: I0202 15:43:20.458224 4733 scope.go:117] "RemoveContainer" containerID="aa3ebde02e5a5d6bc82ad8bf39facfc176dfcc69833c63a12ba398b12435f40e" Feb 02 15:43:20 crc kubenswrapper[4733]: I0202 15:43:20.492915 4733 scope.go:117] "RemoveContainer" containerID="4f7eae9add368d7a939b590643cc6b2b137344a34ce936373e07f02c65d9fd86" Feb 02 15:43:20 crc kubenswrapper[4733]: I0202 15:43:20.508866 4733 scope.go:117] "RemoveContainer" containerID="1d64e3457cc83f08629d107d40db515c344d361ad519db210f6a4ea6c37f5b37" Feb 02 15:43:20 crc kubenswrapper[4733]: I0202 15:43:20.549059 4733 scope.go:117] "RemoveContainer" containerID="1093dfe86ecb94863ae3a05142c42aa595e42408f929d5bf60be16907a8a2fd3" Feb 02 15:43:34 crc kubenswrapper[4733]: I0202 15:43:34.988004 4733 patch_prober.go:28] interesting pod/machine-config-daemon-zgsv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 15:43:34 crc kubenswrapper[4733]: I0202 15:43:34.988551 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 02 15:43:34 crc kubenswrapper[4733]: I0202 15:43:34.988605 4733 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" Feb 02 15:43:34 crc kubenswrapper[4733]: I0202 15:43:34.989260 4733 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8837c504005a41d8189bc30082d6c198635f60c659df3e0066907bf203ae6b97"} pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 02 15:43:34 crc kubenswrapper[4733]: I0202 15:43:34.989312 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" containerName="machine-config-daemon" containerID="cri-o://8837c504005a41d8189bc30082d6c198635f60c659df3e0066907bf203ae6b97" gracePeriod=600 Feb 02 15:43:35 crc kubenswrapper[4733]: E0202 15:43:35.122928 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zgsv9_openshift-machine-config-operator(83d0e810-ea5c-47aa-947b-7865b045bf94)\"" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" Feb 02 15:43:36 crc kubenswrapper[4733]: I0202 15:43:36.102647 4733 generic.go:334] "Generic (PLEG): container finished" podID="83d0e810-ea5c-47aa-947b-7865b045bf94" containerID="8837c504005a41d8189bc30082d6c198635f60c659df3e0066907bf203ae6b97" exitCode=0 Feb 02 15:43:36 crc kubenswrapper[4733]: I0202 15:43:36.102786 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" event={"ID":"83d0e810-ea5c-47aa-947b-7865b045bf94","Type":"ContainerDied","Data":"8837c504005a41d8189bc30082d6c198635f60c659df3e0066907bf203ae6b97"} Feb 02 15:43:36 crc kubenswrapper[4733]: I0202 15:43:36.103429 4733 scope.go:117] "RemoveContainer" containerID="9ef5e0247a233902273cc6edac5d4e35e73d0f3d5ded3fb8bb539db06e43f773" Feb 02 15:43:36 crc kubenswrapper[4733]: I0202 15:43:36.104770 4733 scope.go:117] "RemoveContainer" containerID="8837c504005a41d8189bc30082d6c198635f60c659df3e0066907bf203ae6b97" Feb 02 15:43:36 crc kubenswrapper[4733]: E0202 15:43:36.105632 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zgsv9_openshift-machine-config-operator(83d0e810-ea5c-47aa-947b-7865b045bf94)\"" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" Feb 02 15:43:48 crc kubenswrapper[4733]: I0202 15:43:48.254676 4733 scope.go:117] "RemoveContainer" containerID="8837c504005a41d8189bc30082d6c198635f60c659df3e0066907bf203ae6b97" Feb 02 15:43:48 crc kubenswrapper[4733]: E0202 15:43:48.255585 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zgsv9_openshift-machine-config-operator(83d0e810-ea5c-47aa-947b-7865b045bf94)\"" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" Feb 02 15:43:49 crc kubenswrapper[4733]: I0202 15:43:49.800434 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-p4b26"] Feb 02 15:43:49 crc kubenswrapper[4733]: E0202 15:43:49.802872 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5893b4e6-a11b-4e61-ba80-2ab3963863db" containerName="gather" Feb 02 15:43:49 crc kubenswrapper[4733]: I0202 15:43:49.803051 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="5893b4e6-a11b-4e61-ba80-2ab3963863db" containerName="gather" Feb 02 15:43:49 crc kubenswrapper[4733]: E0202 15:43:49.803153 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5893b4e6-a11b-4e61-ba80-2ab3963863db" containerName="copy" Feb 02 15:43:49 crc kubenswrapper[4733]: I0202 15:43:49.803292 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="5893b4e6-a11b-4e61-ba80-2ab3963863db" containerName="copy" Feb 02 15:43:49 crc kubenswrapper[4733]: I0202 15:43:49.804619 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="5893b4e6-a11b-4e61-ba80-2ab3963863db" containerName="copy" Feb 02 15:43:49 crc kubenswrapper[4733]: I0202 15:43:49.804879 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="5893b4e6-a11b-4e61-ba80-2ab3963863db" containerName="gather" Feb 02 15:43:49 crc kubenswrapper[4733]: I0202 15:43:49.806342 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p4b26" Feb 02 15:43:49 crc kubenswrapper[4733]: I0202 15:43:49.814239 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p4b26"] Feb 02 15:43:49 crc kubenswrapper[4733]: I0202 15:43:49.962966 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/440d41ac-795f-49ce-b826-0032f6578a0b-catalog-content\") pod \"certified-operators-p4b26\" (UID: \"440d41ac-795f-49ce-b826-0032f6578a0b\") " pod="openshift-marketplace/certified-operators-p4b26" Feb 02 15:43:49 crc kubenswrapper[4733]: I0202 15:43:49.963047 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/440d41ac-795f-49ce-b826-0032f6578a0b-utilities\") pod \"certified-operators-p4b26\" (UID: \"440d41ac-795f-49ce-b826-0032f6578a0b\") " pod="openshift-marketplace/certified-operators-p4b26" Feb 02 15:43:49 crc kubenswrapper[4733]: I0202 15:43:49.963104 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5slkc\" (UniqueName: \"kubernetes.io/projected/440d41ac-795f-49ce-b826-0032f6578a0b-kube-api-access-5slkc\") pod \"certified-operators-p4b26\" (UID: \"440d41ac-795f-49ce-b826-0032f6578a0b\") " pod="openshift-marketplace/certified-operators-p4b26" Feb 02 15:43:50 crc kubenswrapper[4733]: I0202 15:43:50.063734 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5slkc\" (UniqueName: \"kubernetes.io/projected/440d41ac-795f-49ce-b826-0032f6578a0b-kube-api-access-5slkc\") pod \"certified-operators-p4b26\" (UID: \"440d41ac-795f-49ce-b826-0032f6578a0b\") " pod="openshift-marketplace/certified-operators-p4b26" Feb 02 15:43:50 crc kubenswrapper[4733]: I0202 15:43:50.063831 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/440d41ac-795f-49ce-b826-0032f6578a0b-catalog-content\") pod \"certified-operators-p4b26\" (UID: \"440d41ac-795f-49ce-b826-0032f6578a0b\") " pod="openshift-marketplace/certified-operators-p4b26" Feb 02 15:43:50 crc kubenswrapper[4733]: I0202 15:43:50.063878 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/440d41ac-795f-49ce-b826-0032f6578a0b-utilities\") pod \"certified-operators-p4b26\" (UID: \"440d41ac-795f-49ce-b826-0032f6578a0b\") " pod="openshift-marketplace/certified-operators-p4b26" Feb 02 15:43:50 crc kubenswrapper[4733]: I0202 15:43:50.064340 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/440d41ac-795f-49ce-b826-0032f6578a0b-utilities\") pod \"certified-operators-p4b26\" (UID: \"440d41ac-795f-49ce-b826-0032f6578a0b\") " pod="openshift-marketplace/certified-operators-p4b26" Feb 02 15:43:50 crc kubenswrapper[4733]: I0202 15:43:50.064482 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/440d41ac-795f-49ce-b826-0032f6578a0b-catalog-content\") pod \"certified-operators-p4b26\" (UID: \"440d41ac-795f-49ce-b826-0032f6578a0b\") " pod="openshift-marketplace/certified-operators-p4b26" Feb 02 15:43:50 crc kubenswrapper[4733]: I0202 15:43:50.085071 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5slkc\" (UniqueName: \"kubernetes.io/projected/440d41ac-795f-49ce-b826-0032f6578a0b-kube-api-access-5slkc\") pod \"certified-operators-p4b26\" (UID: \"440d41ac-795f-49ce-b826-0032f6578a0b\") " pod="openshift-marketplace/certified-operators-p4b26" Feb 02 15:43:50 crc kubenswrapper[4733]: I0202 15:43:50.129246 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p4b26" Feb 02 15:43:50 crc kubenswrapper[4733]: I0202 15:43:50.587388 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p4b26"] Feb 02 15:43:51 crc kubenswrapper[4733]: I0202 15:43:51.220407 4733 generic.go:334] "Generic (PLEG): container finished" podID="440d41ac-795f-49ce-b826-0032f6578a0b" containerID="48596d1599901659f39ef7d0c264692143c036031f7040fc961a648c66f9dbb9" exitCode=0 Feb 02 15:43:51 crc kubenswrapper[4733]: I0202 15:43:51.220475 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p4b26" event={"ID":"440d41ac-795f-49ce-b826-0032f6578a0b","Type":"ContainerDied","Data":"48596d1599901659f39ef7d0c264692143c036031f7040fc961a648c66f9dbb9"} Feb 02 15:43:51 crc kubenswrapper[4733]: I0202 15:43:51.220515 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p4b26" event={"ID":"440d41ac-795f-49ce-b826-0032f6578a0b","Type":"ContainerStarted","Data":"a48e1ecdb8e29f4d3d584c0c8d577515f630518997b1a1e062ccb8e5efed6d0e"} Feb 02 15:43:51 crc kubenswrapper[4733]: I0202 15:43:51.224782 4733 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 02 15:43:53 crc kubenswrapper[4733]: I0202 15:43:53.234122 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p4b26" event={"ID":"440d41ac-795f-49ce-b826-0032f6578a0b","Type":"ContainerStarted","Data":"d95bdde25ff6e221e2478a1b750a9c7197508ee57ca1b18d138ef6ca3c00208c"} Feb 02 15:43:54 crc kubenswrapper[4733]: I0202 15:43:54.245033 4733 generic.go:334] "Generic (PLEG): container finished" podID="440d41ac-795f-49ce-b826-0032f6578a0b" containerID="d95bdde25ff6e221e2478a1b750a9c7197508ee57ca1b18d138ef6ca3c00208c" exitCode=0 Feb 02 15:43:54 crc kubenswrapper[4733]: I0202 15:43:54.245104 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p4b26" event={"ID":"440d41ac-795f-49ce-b826-0032f6578a0b","Type":"ContainerDied","Data":"d95bdde25ff6e221e2478a1b750a9c7197508ee57ca1b18d138ef6ca3c00208c"} Feb 02 15:43:55 crc kubenswrapper[4733]: I0202 15:43:55.253833 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p4b26" event={"ID":"440d41ac-795f-49ce-b826-0032f6578a0b","Type":"ContainerStarted","Data":"b8db6ce80871e4a44d4b10648a2b153ceb256f23cb2d666fa79ae66f6f7d399c"} Feb 02 15:43:55 crc kubenswrapper[4733]: I0202 15:43:55.282582 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-p4b26" podStartSLOduration=2.842843014 podStartE2EDuration="6.282562065s" podCreationTimestamp="2026-02-02 15:43:49 +0000 UTC" firstStartedPulling="2026-02-02 15:43:51.224346931 +0000 UTC m=+1714.675808329" lastFinishedPulling="2026-02-02 15:43:54.664065982 +0000 UTC m=+1718.115527380" observedRunningTime="2026-02-02 15:43:55.275639431 +0000 UTC m=+1718.727100889" watchObservedRunningTime="2026-02-02 15:43:55.282562065 +0000 UTC m=+1718.734023423" Feb 02 15:43:59 crc kubenswrapper[4733]: I0202 15:43:59.255704 4733 scope.go:117] "RemoveContainer" containerID="8837c504005a41d8189bc30082d6c198635f60c659df3e0066907bf203ae6b97" Feb 02 15:43:59 crc kubenswrapper[4733]: E0202 15:43:59.256385 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zgsv9_openshift-machine-config-operator(83d0e810-ea5c-47aa-947b-7865b045bf94)\"" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" Feb 02 15:44:00 crc kubenswrapper[4733]: I0202 15:44:00.129608 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-p4b26" Feb 02 15:44:00 crc kubenswrapper[4733]: I0202 15:44:00.129678 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-p4b26" Feb 02 15:44:00 crc kubenswrapper[4733]: I0202 15:44:00.204713 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-p4b26" Feb 02 15:44:00 crc kubenswrapper[4733]: I0202 15:44:00.366340 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-p4b26" Feb 02 15:44:00 crc kubenswrapper[4733]: I0202 15:44:00.439211 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p4b26"] Feb 02 15:44:02 crc kubenswrapper[4733]: I0202 15:44:02.312806 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-p4b26" podUID="440d41ac-795f-49ce-b826-0032f6578a0b" containerName="registry-server" containerID="cri-o://b8db6ce80871e4a44d4b10648a2b153ceb256f23cb2d666fa79ae66f6f7d399c" gracePeriod=2 Feb 02 15:44:02 crc kubenswrapper[4733]: I0202 15:44:02.780542 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p4b26" Feb 02 15:44:02 crc kubenswrapper[4733]: I0202 15:44:02.952662 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/440d41ac-795f-49ce-b826-0032f6578a0b-utilities\") pod \"440d41ac-795f-49ce-b826-0032f6578a0b\" (UID: \"440d41ac-795f-49ce-b826-0032f6578a0b\") " Feb 02 15:44:02 crc kubenswrapper[4733]: I0202 15:44:02.952749 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/440d41ac-795f-49ce-b826-0032f6578a0b-catalog-content\") pod \"440d41ac-795f-49ce-b826-0032f6578a0b\" (UID: \"440d41ac-795f-49ce-b826-0032f6578a0b\") " Feb 02 15:44:02 crc kubenswrapper[4733]: I0202 15:44:02.952850 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5slkc\" (UniqueName: \"kubernetes.io/projected/440d41ac-795f-49ce-b826-0032f6578a0b-kube-api-access-5slkc\") pod \"440d41ac-795f-49ce-b826-0032f6578a0b\" (UID: \"440d41ac-795f-49ce-b826-0032f6578a0b\") " Feb 02 15:44:02 crc kubenswrapper[4733]: I0202 15:44:02.954036 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/440d41ac-795f-49ce-b826-0032f6578a0b-utilities" (OuterVolumeSpecName: "utilities") pod "440d41ac-795f-49ce-b826-0032f6578a0b" (UID: "440d41ac-795f-49ce-b826-0032f6578a0b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:44:02 crc kubenswrapper[4733]: I0202 15:44:02.964084 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/440d41ac-795f-49ce-b826-0032f6578a0b-kube-api-access-5slkc" (OuterVolumeSpecName: "kube-api-access-5slkc") pod "440d41ac-795f-49ce-b826-0032f6578a0b" (UID: "440d41ac-795f-49ce-b826-0032f6578a0b"). InnerVolumeSpecName "kube-api-access-5slkc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:44:03 crc kubenswrapper[4733]: I0202 15:44:03.030756 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/440d41ac-795f-49ce-b826-0032f6578a0b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "440d41ac-795f-49ce-b826-0032f6578a0b" (UID: "440d41ac-795f-49ce-b826-0032f6578a0b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:44:03 crc kubenswrapper[4733]: I0202 15:44:03.054746 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5slkc\" (UniqueName: \"kubernetes.io/projected/440d41ac-795f-49ce-b826-0032f6578a0b-kube-api-access-5slkc\") on node \"crc\" DevicePath \"\"" Feb 02 15:44:03 crc kubenswrapper[4733]: I0202 15:44:03.054792 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/440d41ac-795f-49ce-b826-0032f6578a0b-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 15:44:03 crc kubenswrapper[4733]: I0202 15:44:03.054824 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/440d41ac-795f-49ce-b826-0032f6578a0b-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 15:44:03 crc kubenswrapper[4733]: I0202 15:44:03.322041 4733 generic.go:334] "Generic (PLEG): container finished" podID="440d41ac-795f-49ce-b826-0032f6578a0b" containerID="b8db6ce80871e4a44d4b10648a2b153ceb256f23cb2d666fa79ae66f6f7d399c" exitCode=0 Feb 02 15:44:03 crc kubenswrapper[4733]: I0202 15:44:03.322103 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p4b26" event={"ID":"440d41ac-795f-49ce-b826-0032f6578a0b","Type":"ContainerDied","Data":"b8db6ce80871e4a44d4b10648a2b153ceb256f23cb2d666fa79ae66f6f7d399c"} Feb 02 15:44:03 crc kubenswrapper[4733]: I0202 15:44:03.322463 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p4b26" event={"ID":"440d41ac-795f-49ce-b826-0032f6578a0b","Type":"ContainerDied","Data":"a48e1ecdb8e29f4d3d584c0c8d577515f630518997b1a1e062ccb8e5efed6d0e"} Feb 02 15:44:03 crc kubenswrapper[4733]: I0202 15:44:03.322500 4733 scope.go:117] "RemoveContainer" containerID="b8db6ce80871e4a44d4b10648a2b153ceb256f23cb2d666fa79ae66f6f7d399c" Feb 02 15:44:03 crc kubenswrapper[4733]: I0202 15:44:03.322142 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p4b26" Feb 02 15:44:03 crc kubenswrapper[4733]: I0202 15:44:03.350684 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p4b26"] Feb 02 15:44:03 crc kubenswrapper[4733]: I0202 15:44:03.355030 4733 scope.go:117] "RemoveContainer" containerID="d95bdde25ff6e221e2478a1b750a9c7197508ee57ca1b18d138ef6ca3c00208c" Feb 02 15:44:03 crc kubenswrapper[4733]: I0202 15:44:03.356062 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-p4b26"] Feb 02 15:44:03 crc kubenswrapper[4733]: I0202 15:44:03.374537 4733 scope.go:117] "RemoveContainer" containerID="48596d1599901659f39ef7d0c264692143c036031f7040fc961a648c66f9dbb9" Feb 02 15:44:03 crc kubenswrapper[4733]: I0202 15:44:03.397804 4733 scope.go:117] "RemoveContainer" containerID="b8db6ce80871e4a44d4b10648a2b153ceb256f23cb2d666fa79ae66f6f7d399c" Feb 02 15:44:03 crc kubenswrapper[4733]: E0202 15:44:03.398561 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8db6ce80871e4a44d4b10648a2b153ceb256f23cb2d666fa79ae66f6f7d399c\": container with ID starting with b8db6ce80871e4a44d4b10648a2b153ceb256f23cb2d666fa79ae66f6f7d399c not found: ID does not exist" containerID="b8db6ce80871e4a44d4b10648a2b153ceb256f23cb2d666fa79ae66f6f7d399c" Feb 02 15:44:03 crc kubenswrapper[4733]: I0202 15:44:03.398612 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8db6ce80871e4a44d4b10648a2b153ceb256f23cb2d666fa79ae66f6f7d399c"} err="failed to get container status \"b8db6ce80871e4a44d4b10648a2b153ceb256f23cb2d666fa79ae66f6f7d399c\": rpc error: code = NotFound desc = could not find container \"b8db6ce80871e4a44d4b10648a2b153ceb256f23cb2d666fa79ae66f6f7d399c\": container with ID starting with b8db6ce80871e4a44d4b10648a2b153ceb256f23cb2d666fa79ae66f6f7d399c not found: ID does not exist" Feb 02 15:44:03 crc kubenswrapper[4733]: I0202 15:44:03.398641 4733 scope.go:117] "RemoveContainer" containerID="d95bdde25ff6e221e2478a1b750a9c7197508ee57ca1b18d138ef6ca3c00208c" Feb 02 15:44:03 crc kubenswrapper[4733]: E0202 15:44:03.398940 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d95bdde25ff6e221e2478a1b750a9c7197508ee57ca1b18d138ef6ca3c00208c\": container with ID starting with d95bdde25ff6e221e2478a1b750a9c7197508ee57ca1b18d138ef6ca3c00208c not found: ID does not exist" containerID="d95bdde25ff6e221e2478a1b750a9c7197508ee57ca1b18d138ef6ca3c00208c" Feb 02 15:44:03 crc kubenswrapper[4733]: I0202 15:44:03.398978 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d95bdde25ff6e221e2478a1b750a9c7197508ee57ca1b18d138ef6ca3c00208c"} err="failed to get container status \"d95bdde25ff6e221e2478a1b750a9c7197508ee57ca1b18d138ef6ca3c00208c\": rpc error: code = NotFound desc = could not find container \"d95bdde25ff6e221e2478a1b750a9c7197508ee57ca1b18d138ef6ca3c00208c\": container with ID starting with d95bdde25ff6e221e2478a1b750a9c7197508ee57ca1b18d138ef6ca3c00208c not found: ID does not exist" Feb 02 15:44:03 crc kubenswrapper[4733]: I0202 15:44:03.399002 4733 scope.go:117] "RemoveContainer" containerID="48596d1599901659f39ef7d0c264692143c036031f7040fc961a648c66f9dbb9" Feb 02 15:44:03 crc kubenswrapper[4733]: E0202 15:44:03.399533 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48596d1599901659f39ef7d0c264692143c036031f7040fc961a648c66f9dbb9\": container with ID starting with 48596d1599901659f39ef7d0c264692143c036031f7040fc961a648c66f9dbb9 not found: ID does not exist" containerID="48596d1599901659f39ef7d0c264692143c036031f7040fc961a648c66f9dbb9" Feb 02 15:44:03 crc kubenswrapper[4733]: I0202 15:44:03.399581 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48596d1599901659f39ef7d0c264692143c036031f7040fc961a648c66f9dbb9"} err="failed to get container status \"48596d1599901659f39ef7d0c264692143c036031f7040fc961a648c66f9dbb9\": rpc error: code = NotFound desc = could not find container \"48596d1599901659f39ef7d0c264692143c036031f7040fc961a648c66f9dbb9\": container with ID starting with 48596d1599901659f39ef7d0c264692143c036031f7040fc961a648c66f9dbb9 not found: ID does not exist" Feb 02 15:44:05 crc kubenswrapper[4733]: I0202 15:44:05.264606 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="440d41ac-795f-49ce-b826-0032f6578a0b" path="/var/lib/kubelet/pods/440d41ac-795f-49ce-b826-0032f6578a0b/volumes" Feb 02 15:44:12 crc kubenswrapper[4733]: I0202 15:44:12.255838 4733 scope.go:117] "RemoveContainer" containerID="8837c504005a41d8189bc30082d6c198635f60c659df3e0066907bf203ae6b97" Feb 02 15:44:12 crc kubenswrapper[4733]: E0202 15:44:12.256596 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zgsv9_openshift-machine-config-operator(83d0e810-ea5c-47aa-947b-7865b045bf94)\"" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" Feb 02 15:44:20 crc kubenswrapper[4733]: I0202 15:44:20.621819 4733 scope.go:117] "RemoveContainer" containerID="1330407f99c2108f44a26a135750204597d0a5b48f21a356bb3feb3388cf237b" Feb 02 15:44:20 crc kubenswrapper[4733]: I0202 15:44:20.688230 4733 scope.go:117] "RemoveContainer" containerID="461dca89c3c6f78c21b9d2e862cb3b9cef70af44ccef57b4d7e23e7dfa2cde43" Feb 02 15:44:20 crc kubenswrapper[4733]: I0202 15:44:20.713620 4733 scope.go:117] "RemoveContainer" containerID="7909cf6a610bf1ebc516cb20c801752e47646be453177b9151b29750fcba7c66" Feb 02 15:44:20 crc kubenswrapper[4733]: I0202 15:44:20.750443 4733 scope.go:117] "RemoveContainer" containerID="57cd1e3b7613fed348ef3bc76d834b92b9b0a397c1157db8837ddab596c5bcad" Feb 02 15:44:25 crc kubenswrapper[4733]: I0202 15:44:25.255286 4733 scope.go:117] "RemoveContainer" containerID="8837c504005a41d8189bc30082d6c198635f60c659df3e0066907bf203ae6b97" Feb 02 15:44:25 crc kubenswrapper[4733]: E0202 15:44:25.255644 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zgsv9_openshift-machine-config-operator(83d0e810-ea5c-47aa-947b-7865b045bf94)\"" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" Feb 02 15:44:36 crc kubenswrapper[4733]: I0202 15:44:36.255613 4733 scope.go:117] "RemoveContainer" containerID="8837c504005a41d8189bc30082d6c198635f60c659df3e0066907bf203ae6b97" Feb 02 15:44:36 crc kubenswrapper[4733]: E0202 15:44:36.256627 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zgsv9_openshift-machine-config-operator(83d0e810-ea5c-47aa-947b-7865b045bf94)\"" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" Feb 02 15:44:48 crc kubenswrapper[4733]: I0202 15:44:48.255234 4733 scope.go:117] "RemoveContainer" containerID="8837c504005a41d8189bc30082d6c198635f60c659df3e0066907bf203ae6b97" Feb 02 15:44:48 crc kubenswrapper[4733]: E0202 15:44:48.256267 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zgsv9_openshift-machine-config-operator(83d0e810-ea5c-47aa-947b-7865b045bf94)\"" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" Feb 02 15:45:00 crc kubenswrapper[4733]: I0202 15:45:00.157537 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500785-v6cdw"] Feb 02 15:45:00 crc kubenswrapper[4733]: E0202 15:45:00.158494 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="440d41ac-795f-49ce-b826-0032f6578a0b" containerName="extract-content" Feb 02 15:45:00 crc kubenswrapper[4733]: I0202 15:45:00.158511 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="440d41ac-795f-49ce-b826-0032f6578a0b" containerName="extract-content" Feb 02 15:45:00 crc kubenswrapper[4733]: E0202 15:45:00.158535 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="440d41ac-795f-49ce-b826-0032f6578a0b" containerName="registry-server" Feb 02 15:45:00 crc kubenswrapper[4733]: I0202 15:45:00.158543 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="440d41ac-795f-49ce-b826-0032f6578a0b" containerName="registry-server" Feb 02 15:45:00 crc kubenswrapper[4733]: E0202 15:45:00.158558 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="440d41ac-795f-49ce-b826-0032f6578a0b" containerName="extract-utilities" Feb 02 15:45:00 crc kubenswrapper[4733]: I0202 15:45:00.158568 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="440d41ac-795f-49ce-b826-0032f6578a0b" containerName="extract-utilities" Feb 02 15:45:00 crc kubenswrapper[4733]: I0202 15:45:00.158727 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="440d41ac-795f-49ce-b826-0032f6578a0b" containerName="registry-server" Feb 02 15:45:00 crc kubenswrapper[4733]: I0202 15:45:00.159617 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500785-v6cdw" Feb 02 15:45:00 crc kubenswrapper[4733]: I0202 15:45:00.162645 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 02 15:45:00 crc kubenswrapper[4733]: I0202 15:45:00.163474 4733 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 02 15:45:00 crc kubenswrapper[4733]: I0202 15:45:00.172417 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500785-v6cdw"] Feb 02 15:45:00 crc kubenswrapper[4733]: I0202 15:45:00.251529 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/381b7cac-435b-41cb-9588-3b759e43d0ee-config-volume\") pod \"collect-profiles-29500785-v6cdw\" (UID: \"381b7cac-435b-41cb-9588-3b759e43d0ee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500785-v6cdw" Feb 02 15:45:00 crc kubenswrapper[4733]: I0202 15:45:00.251956 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtwdc\" (UniqueName: \"kubernetes.io/projected/381b7cac-435b-41cb-9588-3b759e43d0ee-kube-api-access-dtwdc\") pod \"collect-profiles-29500785-v6cdw\" (UID: \"381b7cac-435b-41cb-9588-3b759e43d0ee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500785-v6cdw" Feb 02 15:45:00 crc kubenswrapper[4733]: I0202 15:45:00.252114 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/381b7cac-435b-41cb-9588-3b759e43d0ee-secret-volume\") pod \"collect-profiles-29500785-v6cdw\" (UID: \"381b7cac-435b-41cb-9588-3b759e43d0ee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500785-v6cdw" Feb 02 15:45:00 crc kubenswrapper[4733]: I0202 15:45:00.353488 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtwdc\" (UniqueName: \"kubernetes.io/projected/381b7cac-435b-41cb-9588-3b759e43d0ee-kube-api-access-dtwdc\") pod \"collect-profiles-29500785-v6cdw\" (UID: \"381b7cac-435b-41cb-9588-3b759e43d0ee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500785-v6cdw" Feb 02 15:45:00 crc kubenswrapper[4733]: I0202 15:45:00.353550 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/381b7cac-435b-41cb-9588-3b759e43d0ee-secret-volume\") pod \"collect-profiles-29500785-v6cdw\" (UID: \"381b7cac-435b-41cb-9588-3b759e43d0ee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500785-v6cdw" Feb 02 15:45:00 crc kubenswrapper[4733]: I0202 15:45:00.353619 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/381b7cac-435b-41cb-9588-3b759e43d0ee-config-volume\") pod \"collect-profiles-29500785-v6cdw\" (UID: \"381b7cac-435b-41cb-9588-3b759e43d0ee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500785-v6cdw" Feb 02 15:45:00 crc kubenswrapper[4733]: I0202 15:45:00.354748 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/381b7cac-435b-41cb-9588-3b759e43d0ee-config-volume\") pod \"collect-profiles-29500785-v6cdw\" (UID: \"381b7cac-435b-41cb-9588-3b759e43d0ee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500785-v6cdw" Feb 02 15:45:00 crc kubenswrapper[4733]: I0202 15:45:00.361625 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/381b7cac-435b-41cb-9588-3b759e43d0ee-secret-volume\") pod \"collect-profiles-29500785-v6cdw\" (UID: \"381b7cac-435b-41cb-9588-3b759e43d0ee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500785-v6cdw" Feb 02 15:45:00 crc kubenswrapper[4733]: I0202 15:45:00.376619 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtwdc\" (UniqueName: \"kubernetes.io/projected/381b7cac-435b-41cb-9588-3b759e43d0ee-kube-api-access-dtwdc\") pod \"collect-profiles-29500785-v6cdw\" (UID: \"381b7cac-435b-41cb-9588-3b759e43d0ee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29500785-v6cdw" Feb 02 15:45:00 crc kubenswrapper[4733]: I0202 15:45:00.486080 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500785-v6cdw" Feb 02 15:45:00 crc kubenswrapper[4733]: I0202 15:45:00.880314 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29500785-v6cdw"] Feb 02 15:45:01 crc kubenswrapper[4733]: I0202 15:45:01.255268 4733 scope.go:117] "RemoveContainer" containerID="8837c504005a41d8189bc30082d6c198635f60c659df3e0066907bf203ae6b97" Feb 02 15:45:01 crc kubenswrapper[4733]: E0202 15:45:01.255489 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zgsv9_openshift-machine-config-operator(83d0e810-ea5c-47aa-947b-7865b045bf94)\"" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" Feb 02 15:45:01 crc kubenswrapper[4733]: I0202 15:45:01.784643 4733 generic.go:334] "Generic (PLEG): container finished" podID="381b7cac-435b-41cb-9588-3b759e43d0ee" containerID="fc45adbf61769816de6f02d99ae32dc98b7aca87b9d186fc2dbf71225613d9c2" exitCode=0 Feb 02 15:45:01 crc kubenswrapper[4733]: I0202 15:45:01.784877 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500785-v6cdw" event={"ID":"381b7cac-435b-41cb-9588-3b759e43d0ee","Type":"ContainerDied","Data":"fc45adbf61769816de6f02d99ae32dc98b7aca87b9d186fc2dbf71225613d9c2"} Feb 02 15:45:01 crc kubenswrapper[4733]: I0202 15:45:01.785128 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500785-v6cdw" event={"ID":"381b7cac-435b-41cb-9588-3b759e43d0ee","Type":"ContainerStarted","Data":"bbfeb786e79f597bb6a6677f07930e92428ac7b255e912f375fba089bb10f4ad"} Feb 02 15:45:03 crc kubenswrapper[4733]: I0202 15:45:03.090223 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500785-v6cdw" Feb 02 15:45:03 crc kubenswrapper[4733]: I0202 15:45:03.098799 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/381b7cac-435b-41cb-9588-3b759e43d0ee-secret-volume\") pod \"381b7cac-435b-41cb-9588-3b759e43d0ee\" (UID: \"381b7cac-435b-41cb-9588-3b759e43d0ee\") " Feb 02 15:45:03 crc kubenswrapper[4733]: I0202 15:45:03.099120 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/381b7cac-435b-41cb-9588-3b759e43d0ee-config-volume\") pod \"381b7cac-435b-41cb-9588-3b759e43d0ee\" (UID: \"381b7cac-435b-41cb-9588-3b759e43d0ee\") " Feb 02 15:45:03 crc kubenswrapper[4733]: I0202 15:45:03.099269 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dtwdc\" (UniqueName: \"kubernetes.io/projected/381b7cac-435b-41cb-9588-3b759e43d0ee-kube-api-access-dtwdc\") pod \"381b7cac-435b-41cb-9588-3b759e43d0ee\" (UID: \"381b7cac-435b-41cb-9588-3b759e43d0ee\") " Feb 02 15:45:03 crc kubenswrapper[4733]: I0202 15:45:03.099848 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/381b7cac-435b-41cb-9588-3b759e43d0ee-config-volume" (OuterVolumeSpecName: "config-volume") pod "381b7cac-435b-41cb-9588-3b759e43d0ee" (UID: "381b7cac-435b-41cb-9588-3b759e43d0ee"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 02 15:45:03 crc kubenswrapper[4733]: I0202 15:45:03.104326 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/381b7cac-435b-41cb-9588-3b759e43d0ee-kube-api-access-dtwdc" (OuterVolumeSpecName: "kube-api-access-dtwdc") pod "381b7cac-435b-41cb-9588-3b759e43d0ee" (UID: "381b7cac-435b-41cb-9588-3b759e43d0ee"). InnerVolumeSpecName "kube-api-access-dtwdc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:45:03 crc kubenswrapper[4733]: I0202 15:45:03.107667 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/381b7cac-435b-41cb-9588-3b759e43d0ee-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "381b7cac-435b-41cb-9588-3b759e43d0ee" (UID: "381b7cac-435b-41cb-9588-3b759e43d0ee"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 02 15:45:03 crc kubenswrapper[4733]: I0202 15:45:03.201682 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dtwdc\" (UniqueName: \"kubernetes.io/projected/381b7cac-435b-41cb-9588-3b759e43d0ee-kube-api-access-dtwdc\") on node \"crc\" DevicePath \"\"" Feb 02 15:45:03 crc kubenswrapper[4733]: I0202 15:45:03.201740 4733 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/381b7cac-435b-41cb-9588-3b759e43d0ee-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 02 15:45:03 crc kubenswrapper[4733]: I0202 15:45:03.201758 4733 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/381b7cac-435b-41cb-9588-3b759e43d0ee-config-volume\") on node \"crc\" DevicePath \"\"" Feb 02 15:45:03 crc kubenswrapper[4733]: I0202 15:45:03.800914 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29500785-v6cdw" event={"ID":"381b7cac-435b-41cb-9588-3b759e43d0ee","Type":"ContainerDied","Data":"bbfeb786e79f597bb6a6677f07930e92428ac7b255e912f375fba089bb10f4ad"} Feb 02 15:45:03 crc kubenswrapper[4733]: I0202 15:45:03.800968 4733 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bbfeb786e79f597bb6a6677f07930e92428ac7b255e912f375fba089bb10f4ad" Feb 02 15:45:03 crc kubenswrapper[4733]: I0202 15:45:03.801043 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29500785-v6cdw" Feb 02 15:45:16 crc kubenswrapper[4733]: I0202 15:45:16.255767 4733 scope.go:117] "RemoveContainer" containerID="8837c504005a41d8189bc30082d6c198635f60c659df3e0066907bf203ae6b97" Feb 02 15:45:16 crc kubenswrapper[4733]: E0202 15:45:16.256983 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zgsv9_openshift-machine-config-operator(83d0e810-ea5c-47aa-947b-7865b045bf94)\"" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" Feb 02 15:45:19 crc kubenswrapper[4733]: I0202 15:45:19.966407 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-jbls7/must-gather-jwwdr"] Feb 02 15:45:19 crc kubenswrapper[4733]: E0202 15:45:19.967126 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="381b7cac-435b-41cb-9588-3b759e43d0ee" containerName="collect-profiles" Feb 02 15:45:19 crc kubenswrapper[4733]: I0202 15:45:19.967146 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="381b7cac-435b-41cb-9588-3b759e43d0ee" containerName="collect-profiles" Feb 02 15:45:19 crc kubenswrapper[4733]: I0202 15:45:19.967389 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="381b7cac-435b-41cb-9588-3b759e43d0ee" containerName="collect-profiles" Feb 02 15:45:19 crc kubenswrapper[4733]: I0202 15:45:19.969706 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jbls7/must-gather-jwwdr" Feb 02 15:45:19 crc kubenswrapper[4733]: I0202 15:45:19.973305 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-jbls7"/"openshift-service-ca.crt" Feb 02 15:45:19 crc kubenswrapper[4733]: I0202 15:45:19.973821 4733 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-jbls7"/"kube-root-ca.crt" Feb 02 15:45:19 crc kubenswrapper[4733]: I0202 15:45:19.995357 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-jbls7/must-gather-jwwdr"] Feb 02 15:45:20 crc kubenswrapper[4733]: I0202 15:45:20.070434 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/fdb2080f-f5b6-4f5f-9aeb-44b4efaebca6-must-gather-output\") pod \"must-gather-jwwdr\" (UID: \"fdb2080f-f5b6-4f5f-9aeb-44b4efaebca6\") " pod="openshift-must-gather-jbls7/must-gather-jwwdr" Feb 02 15:45:20 crc kubenswrapper[4733]: I0202 15:45:20.070540 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gclnp\" (UniqueName: \"kubernetes.io/projected/fdb2080f-f5b6-4f5f-9aeb-44b4efaebca6-kube-api-access-gclnp\") pod \"must-gather-jwwdr\" (UID: \"fdb2080f-f5b6-4f5f-9aeb-44b4efaebca6\") " pod="openshift-must-gather-jbls7/must-gather-jwwdr" Feb 02 15:45:20 crc kubenswrapper[4733]: I0202 15:45:20.172119 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gclnp\" (UniqueName: \"kubernetes.io/projected/fdb2080f-f5b6-4f5f-9aeb-44b4efaebca6-kube-api-access-gclnp\") pod \"must-gather-jwwdr\" (UID: \"fdb2080f-f5b6-4f5f-9aeb-44b4efaebca6\") " pod="openshift-must-gather-jbls7/must-gather-jwwdr" Feb 02 15:45:20 crc kubenswrapper[4733]: I0202 15:45:20.172226 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/fdb2080f-f5b6-4f5f-9aeb-44b4efaebca6-must-gather-output\") pod \"must-gather-jwwdr\" (UID: \"fdb2080f-f5b6-4f5f-9aeb-44b4efaebca6\") " pod="openshift-must-gather-jbls7/must-gather-jwwdr" Feb 02 15:45:20 crc kubenswrapper[4733]: I0202 15:45:20.172722 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/fdb2080f-f5b6-4f5f-9aeb-44b4efaebca6-must-gather-output\") pod \"must-gather-jwwdr\" (UID: \"fdb2080f-f5b6-4f5f-9aeb-44b4efaebca6\") " pod="openshift-must-gather-jbls7/must-gather-jwwdr" Feb 02 15:45:20 crc kubenswrapper[4733]: I0202 15:45:20.201719 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gclnp\" (UniqueName: \"kubernetes.io/projected/fdb2080f-f5b6-4f5f-9aeb-44b4efaebca6-kube-api-access-gclnp\") pod \"must-gather-jwwdr\" (UID: \"fdb2080f-f5b6-4f5f-9aeb-44b4efaebca6\") " pod="openshift-must-gather-jbls7/must-gather-jwwdr" Feb 02 15:45:20 crc kubenswrapper[4733]: I0202 15:45:20.289561 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jbls7/must-gather-jwwdr" Feb 02 15:45:20 crc kubenswrapper[4733]: I0202 15:45:20.739474 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-jbls7/must-gather-jwwdr"] Feb 02 15:45:20 crc kubenswrapper[4733]: W0202 15:45:20.747138 4733 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfdb2080f_f5b6_4f5f_9aeb_44b4efaebca6.slice/crio-c874d23d4af185bf1dc622e146a8918f6d7889097f470185ca98ecbd0776d230 WatchSource:0}: Error finding container c874d23d4af185bf1dc622e146a8918f6d7889097f470185ca98ecbd0776d230: Status 404 returned error can't find the container with id c874d23d4af185bf1dc622e146a8918f6d7889097f470185ca98ecbd0776d230 Feb 02 15:45:20 crc kubenswrapper[4733]: I0202 15:45:20.850998 4733 scope.go:117] "RemoveContainer" containerID="620e7108713cf105506971c0510c987a106ae4ec3c66f5219886e20d0bc4dbd1" Feb 02 15:45:20 crc kubenswrapper[4733]: I0202 15:45:20.927911 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jbls7/must-gather-jwwdr" event={"ID":"fdb2080f-f5b6-4f5f-9aeb-44b4efaebca6","Type":"ContainerStarted","Data":"c874d23d4af185bf1dc622e146a8918f6d7889097f470185ca98ecbd0776d230"} Feb 02 15:45:21 crc kubenswrapper[4733]: I0202 15:45:21.937274 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jbls7/must-gather-jwwdr" event={"ID":"fdb2080f-f5b6-4f5f-9aeb-44b4efaebca6","Type":"ContainerStarted","Data":"4e232227b19c2c7fd67d5324582dc813695db0ce368183f2a72fd2991c870915"} Feb 02 15:45:21 crc kubenswrapper[4733]: I0202 15:45:21.937322 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jbls7/must-gather-jwwdr" event={"ID":"fdb2080f-f5b6-4f5f-9aeb-44b4efaebca6","Type":"ContainerStarted","Data":"db9bfa7e886378a285fdbbd3809ecfbed7b05ccb83c467fe287b93437e0af37b"} Feb 02 15:45:21 crc kubenswrapper[4733]: I0202 15:45:21.955902 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-jbls7/must-gather-jwwdr" podStartSLOduration=2.955877268 podStartE2EDuration="2.955877268s" podCreationTimestamp="2026-02-02 15:45:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-02 15:45:21.952132128 +0000 UTC m=+1805.403593536" watchObservedRunningTime="2026-02-02 15:45:21.955877268 +0000 UTC m=+1805.407338636" Feb 02 15:45:29 crc kubenswrapper[4733]: I0202 15:45:29.255637 4733 scope.go:117] "RemoveContainer" containerID="8837c504005a41d8189bc30082d6c198635f60c659df3e0066907bf203ae6b97" Feb 02 15:45:29 crc kubenswrapper[4733]: E0202 15:45:29.256366 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zgsv9_openshift-machine-config-operator(83d0e810-ea5c-47aa-947b-7865b045bf94)\"" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" Feb 02 15:45:44 crc kubenswrapper[4733]: I0202 15:45:44.254774 4733 scope.go:117] "RemoveContainer" containerID="8837c504005a41d8189bc30082d6c198635f60c659df3e0066907bf203ae6b97" Feb 02 15:45:44 crc kubenswrapper[4733]: E0202 15:45:44.255544 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zgsv9_openshift-machine-config-operator(83d0e810-ea5c-47aa-947b-7865b045bf94)\"" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" Feb 02 15:45:56 crc kubenswrapper[4733]: I0202 15:45:56.254743 4733 scope.go:117] "RemoveContainer" containerID="8837c504005a41d8189bc30082d6c198635f60c659df3e0066907bf203ae6b97" Feb 02 15:45:56 crc kubenswrapper[4733]: E0202 15:45:56.255455 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zgsv9_openshift-machine-config-operator(83d0e810-ea5c-47aa-947b-7865b045bf94)\"" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" Feb 02 15:46:09 crc kubenswrapper[4733]: I0202 15:46:09.255473 4733 scope.go:117] "RemoveContainer" containerID="8837c504005a41d8189bc30082d6c198635f60c659df3e0066907bf203ae6b97" Feb 02 15:46:09 crc kubenswrapper[4733]: E0202 15:46:09.256445 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zgsv9_openshift-machine-config-operator(83d0e810-ea5c-47aa-947b-7865b045bf94)\"" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" Feb 02 15:46:10 crc kubenswrapper[4733]: I0202 15:46:10.557026 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-dkktg_bd697754-4524-44ed-be19-18747c0626f5/control-plane-machine-set-operator/0.log" Feb 02 15:46:10 crc kubenswrapper[4733]: I0202 15:46:10.708025 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-74k4s_a3e78085-1f5d-4f0c-84bb-62b787b1fbdd/kube-rbac-proxy/0.log" Feb 02 15:46:10 crc kubenswrapper[4733]: I0202 15:46:10.736842 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-74k4s_a3e78085-1f5d-4f0c-84bb-62b787b1fbdd/machine-api-operator/0.log" Feb 02 15:46:21 crc kubenswrapper[4733]: I0202 15:46:21.255047 4733 scope.go:117] "RemoveContainer" containerID="8837c504005a41d8189bc30082d6c198635f60c659df3e0066907bf203ae6b97" Feb 02 15:46:21 crc kubenswrapper[4733]: E0202 15:46:21.257016 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zgsv9_openshift-machine-config-operator(83d0e810-ea5c-47aa-947b-7865b045bf94)\"" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" Feb 02 15:46:32 crc kubenswrapper[4733]: I0202 15:46:32.255466 4733 scope.go:117] "RemoveContainer" containerID="8837c504005a41d8189bc30082d6c198635f60c659df3e0066907bf203ae6b97" Feb 02 15:46:32 crc kubenswrapper[4733]: E0202 15:46:32.256264 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zgsv9_openshift-machine-config-operator(83d0e810-ea5c-47aa-947b-7865b045bf94)\"" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" Feb 02 15:46:39 crc kubenswrapper[4733]: I0202 15:46:39.494744 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-h9fwf_222b6fa1-d59f-4779-8fdf-72c3fd3ea2cc/kube-rbac-proxy/0.log" Feb 02 15:46:39 crc kubenswrapper[4733]: I0202 15:46:39.505999 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-h9fwf_222b6fa1-d59f-4779-8fdf-72c3fd3ea2cc/controller/0.log" Feb 02 15:46:39 crc kubenswrapper[4733]: I0202 15:46:39.704584 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l6zg6_915914ae-77a5-4bf7-9d71-6207b62d2f50/cp-frr-files/0.log" Feb 02 15:46:39 crc kubenswrapper[4733]: I0202 15:46:39.873104 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l6zg6_915914ae-77a5-4bf7-9d71-6207b62d2f50/cp-reloader/0.log" Feb 02 15:46:39 crc kubenswrapper[4733]: I0202 15:46:39.879367 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l6zg6_915914ae-77a5-4bf7-9d71-6207b62d2f50/cp-frr-files/0.log" Feb 02 15:46:39 crc kubenswrapper[4733]: I0202 15:46:39.911350 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l6zg6_915914ae-77a5-4bf7-9d71-6207b62d2f50/cp-reloader/0.log" Feb 02 15:46:39 crc kubenswrapper[4733]: I0202 15:46:39.928816 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l6zg6_915914ae-77a5-4bf7-9d71-6207b62d2f50/cp-metrics/0.log" Feb 02 15:46:40 crc kubenswrapper[4733]: I0202 15:46:40.051771 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l6zg6_915914ae-77a5-4bf7-9d71-6207b62d2f50/cp-frr-files/0.log" Feb 02 15:46:40 crc kubenswrapper[4733]: I0202 15:46:40.116107 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l6zg6_915914ae-77a5-4bf7-9d71-6207b62d2f50/cp-reloader/0.log" Feb 02 15:46:40 crc kubenswrapper[4733]: I0202 15:46:40.129290 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l6zg6_915914ae-77a5-4bf7-9d71-6207b62d2f50/cp-metrics/0.log" Feb 02 15:46:40 crc kubenswrapper[4733]: I0202 15:46:40.140379 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l6zg6_915914ae-77a5-4bf7-9d71-6207b62d2f50/cp-metrics/0.log" Feb 02 15:46:40 crc kubenswrapper[4733]: I0202 15:46:40.251108 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l6zg6_915914ae-77a5-4bf7-9d71-6207b62d2f50/cp-frr-files/0.log" Feb 02 15:46:40 crc kubenswrapper[4733]: I0202 15:46:40.287468 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l6zg6_915914ae-77a5-4bf7-9d71-6207b62d2f50/cp-reloader/0.log" Feb 02 15:46:40 crc kubenswrapper[4733]: I0202 15:46:40.297703 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l6zg6_915914ae-77a5-4bf7-9d71-6207b62d2f50/cp-metrics/0.log" Feb 02 15:46:40 crc kubenswrapper[4733]: I0202 15:46:40.313883 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l6zg6_915914ae-77a5-4bf7-9d71-6207b62d2f50/controller/0.log" Feb 02 15:46:40 crc kubenswrapper[4733]: I0202 15:46:40.479108 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l6zg6_915914ae-77a5-4bf7-9d71-6207b62d2f50/kube-rbac-proxy/0.log" Feb 02 15:46:40 crc kubenswrapper[4733]: I0202 15:46:40.500921 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l6zg6_915914ae-77a5-4bf7-9d71-6207b62d2f50/frr-metrics/0.log" Feb 02 15:46:40 crc kubenswrapper[4733]: I0202 15:46:40.506811 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l6zg6_915914ae-77a5-4bf7-9d71-6207b62d2f50/kube-rbac-proxy-frr/0.log" Feb 02 15:46:40 crc kubenswrapper[4733]: I0202 15:46:40.709345 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l6zg6_915914ae-77a5-4bf7-9d71-6207b62d2f50/reloader/0.log" Feb 02 15:46:40 crc kubenswrapper[4733]: I0202 15:46:40.742913 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-l6zg6_915914ae-77a5-4bf7-9d71-6207b62d2f50/frr/0.log" Feb 02 15:46:40 crc kubenswrapper[4733]: I0202 15:46:40.751417 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7df86c4f6c-5l6sj_70e35df4-ce17-4181-a64c-d7a70160d0e7/frr-k8s-webhook-server/0.log" Feb 02 15:46:40 crc kubenswrapper[4733]: I0202 15:46:40.960973 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-54b6695bd-mwgx2_2e7a0012-8704-4613-9643-ccb780c9d76d/manager/0.log" Feb 02 15:46:41 crc kubenswrapper[4733]: I0202 15:46:41.018897 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-66bb599b85-wt9zk_ffb4086a-a662-4380-8dd0-c218d36e7c31/webhook-server/0.log" Feb 02 15:46:41 crc kubenswrapper[4733]: I0202 15:46:41.039389 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-7pnkj_1c5ad065-b35b-4784-89b6-72e8dd0a1df5/kube-rbac-proxy/0.log" Feb 02 15:46:41 crc kubenswrapper[4733]: I0202 15:46:41.226020 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-7pnkj_1c5ad065-b35b-4784-89b6-72e8dd0a1df5/speaker/0.log" Feb 02 15:46:43 crc kubenswrapper[4733]: I0202 15:46:43.255608 4733 scope.go:117] "RemoveContainer" containerID="8837c504005a41d8189bc30082d6c198635f60c659df3e0066907bf203ae6b97" Feb 02 15:46:43 crc kubenswrapper[4733]: E0202 15:46:43.256274 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zgsv9_openshift-machine-config-operator(83d0e810-ea5c-47aa-947b-7865b045bf94)\"" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" Feb 02 15:46:57 crc kubenswrapper[4733]: I0202 15:46:57.258761 4733 scope.go:117] "RemoveContainer" containerID="8837c504005a41d8189bc30082d6c198635f60c659df3e0066907bf203ae6b97" Feb 02 15:46:57 crc kubenswrapper[4733]: E0202 15:46:57.260058 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zgsv9_openshift-machine-config-operator(83d0e810-ea5c-47aa-947b-7865b045bf94)\"" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" Feb 02 15:47:06 crc kubenswrapper[4733]: I0202 15:47:06.134937 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg_e8c251f4-787d-49f1-8ec3-32f6c5dff2bb/util/0.log" Feb 02 15:47:06 crc kubenswrapper[4733]: I0202 15:47:06.312523 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg_e8c251f4-787d-49f1-8ec3-32f6c5dff2bb/pull/0.log" Feb 02 15:47:06 crc kubenswrapper[4733]: I0202 15:47:06.330744 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg_e8c251f4-787d-49f1-8ec3-32f6c5dff2bb/util/0.log" Feb 02 15:47:06 crc kubenswrapper[4733]: I0202 15:47:06.367243 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg_e8c251f4-787d-49f1-8ec3-32f6c5dff2bb/pull/0.log" Feb 02 15:47:06 crc kubenswrapper[4733]: I0202 15:47:06.469368 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg_e8c251f4-787d-49f1-8ec3-32f6c5dff2bb/util/0.log" Feb 02 15:47:06 crc kubenswrapper[4733]: I0202 15:47:06.486932 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg_e8c251f4-787d-49f1-8ec3-32f6c5dff2bb/pull/0.log" Feb 02 15:47:06 crc kubenswrapper[4733]: I0202 15:47:06.501094 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dc8rqhg_e8c251f4-787d-49f1-8ec3-32f6c5dff2bb/extract/0.log" Feb 02 15:47:06 crc kubenswrapper[4733]: I0202 15:47:06.650074 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mbh5r_b8fd49b0-ac5d-4b93-8c03-7ac35c150aa2/extract-utilities/0.log" Feb 02 15:47:06 crc kubenswrapper[4733]: I0202 15:47:06.776056 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mbh5r_b8fd49b0-ac5d-4b93-8c03-7ac35c150aa2/extract-utilities/0.log" Feb 02 15:47:06 crc kubenswrapper[4733]: I0202 15:47:06.810130 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mbh5r_b8fd49b0-ac5d-4b93-8c03-7ac35c150aa2/extract-content/0.log" Feb 02 15:47:06 crc kubenswrapper[4733]: I0202 15:47:06.826136 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mbh5r_b8fd49b0-ac5d-4b93-8c03-7ac35c150aa2/extract-content/0.log" Feb 02 15:47:06 crc kubenswrapper[4733]: I0202 15:47:06.952890 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mbh5r_b8fd49b0-ac5d-4b93-8c03-7ac35c150aa2/extract-content/0.log" Feb 02 15:47:06 crc kubenswrapper[4733]: I0202 15:47:06.957706 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mbh5r_b8fd49b0-ac5d-4b93-8c03-7ac35c150aa2/extract-utilities/0.log" Feb 02 15:47:07 crc kubenswrapper[4733]: I0202 15:47:07.172209 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qvwpw_54e05a7f-a312-4455-83ec-be97bf446b68/extract-utilities/0.log" Feb 02 15:47:07 crc kubenswrapper[4733]: I0202 15:47:07.318459 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mbh5r_b8fd49b0-ac5d-4b93-8c03-7ac35c150aa2/registry-server/0.log" Feb 02 15:47:07 crc kubenswrapper[4733]: I0202 15:47:07.339919 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qvwpw_54e05a7f-a312-4455-83ec-be97bf446b68/extract-utilities/0.log" Feb 02 15:47:07 crc kubenswrapper[4733]: I0202 15:47:07.351931 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qvwpw_54e05a7f-a312-4455-83ec-be97bf446b68/extract-content/0.log" Feb 02 15:47:07 crc kubenswrapper[4733]: I0202 15:47:07.387643 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qvwpw_54e05a7f-a312-4455-83ec-be97bf446b68/extract-content/0.log" Feb 02 15:47:07 crc kubenswrapper[4733]: I0202 15:47:07.525247 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qvwpw_54e05a7f-a312-4455-83ec-be97bf446b68/extract-content/0.log" Feb 02 15:47:07 crc kubenswrapper[4733]: I0202 15:47:07.532480 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qvwpw_54e05a7f-a312-4455-83ec-be97bf446b68/extract-utilities/0.log" Feb 02 15:47:07 crc kubenswrapper[4733]: I0202 15:47:07.700077 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-ps6dr_babbc7b3-6e09-43a7-9cbc-70ba7076ab9f/marketplace-operator/0.log" Feb 02 15:47:07 crc kubenswrapper[4733]: I0202 15:47:07.792078 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-smvdc_0367d1cb-4122-4c7e-b203-d5c4216c6ef5/extract-utilities/0.log" Feb 02 15:47:07 crc kubenswrapper[4733]: I0202 15:47:07.902973 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qvwpw_54e05a7f-a312-4455-83ec-be97bf446b68/registry-server/0.log" Feb 02 15:47:07 crc kubenswrapper[4733]: I0202 15:47:07.956079 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-smvdc_0367d1cb-4122-4c7e-b203-d5c4216c6ef5/extract-content/0.log" Feb 02 15:47:07 crc kubenswrapper[4733]: I0202 15:47:07.979285 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-smvdc_0367d1cb-4122-4c7e-b203-d5c4216c6ef5/extract-utilities/0.log" Feb 02 15:47:08 crc kubenswrapper[4733]: I0202 15:47:08.032622 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-smvdc_0367d1cb-4122-4c7e-b203-d5c4216c6ef5/extract-content/0.log" Feb 02 15:47:08 crc kubenswrapper[4733]: I0202 15:47:08.154282 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-smvdc_0367d1cb-4122-4c7e-b203-d5c4216c6ef5/extract-content/0.log" Feb 02 15:47:08 crc kubenswrapper[4733]: I0202 15:47:08.223207 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-smvdc_0367d1cb-4122-4c7e-b203-d5c4216c6ef5/extract-utilities/0.log" Feb 02 15:47:08 crc kubenswrapper[4733]: I0202 15:47:08.300562 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-smvdc_0367d1cb-4122-4c7e-b203-d5c4216c6ef5/registry-server/0.log" Feb 02 15:47:08 crc kubenswrapper[4733]: I0202 15:47:08.353731 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-kqbhk_bcccf11d-2567-42d7-81e2-5684b4b43823/extract-utilities/0.log" Feb 02 15:47:08 crc kubenswrapper[4733]: I0202 15:47:08.506521 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-kqbhk_bcccf11d-2567-42d7-81e2-5684b4b43823/extract-utilities/0.log" Feb 02 15:47:08 crc kubenswrapper[4733]: I0202 15:47:08.510782 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-kqbhk_bcccf11d-2567-42d7-81e2-5684b4b43823/extract-content/0.log" Feb 02 15:47:08 crc kubenswrapper[4733]: I0202 15:47:08.536650 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-kqbhk_bcccf11d-2567-42d7-81e2-5684b4b43823/extract-content/0.log" Feb 02 15:47:08 crc kubenswrapper[4733]: I0202 15:47:08.673154 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-kqbhk_bcccf11d-2567-42d7-81e2-5684b4b43823/extract-content/0.log" Feb 02 15:47:08 crc kubenswrapper[4733]: I0202 15:47:08.686186 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-kqbhk_bcccf11d-2567-42d7-81e2-5684b4b43823/extract-utilities/0.log" Feb 02 15:47:08 crc kubenswrapper[4733]: I0202 15:47:08.930564 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-kqbhk_bcccf11d-2567-42d7-81e2-5684b4b43823/registry-server/0.log" Feb 02 15:47:11 crc kubenswrapper[4733]: I0202 15:47:11.255279 4733 scope.go:117] "RemoveContainer" containerID="8837c504005a41d8189bc30082d6c198635f60c659df3e0066907bf203ae6b97" Feb 02 15:47:11 crc kubenswrapper[4733]: E0202 15:47:11.255897 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zgsv9_openshift-machine-config-operator(83d0e810-ea5c-47aa-947b-7865b045bf94)\"" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" Feb 02 15:47:25 crc kubenswrapper[4733]: I0202 15:47:25.256112 4733 scope.go:117] "RemoveContainer" containerID="8837c504005a41d8189bc30082d6c198635f60c659df3e0066907bf203ae6b97" Feb 02 15:47:25 crc kubenswrapper[4733]: E0202 15:47:25.258331 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zgsv9_openshift-machine-config-operator(83d0e810-ea5c-47aa-947b-7865b045bf94)\"" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" Feb 02 15:47:36 crc kubenswrapper[4733]: I0202 15:47:36.255355 4733 scope.go:117] "RemoveContainer" containerID="8837c504005a41d8189bc30082d6c198635f60c659df3e0066907bf203ae6b97" Feb 02 15:47:36 crc kubenswrapper[4733]: E0202 15:47:36.256074 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zgsv9_openshift-machine-config-operator(83d0e810-ea5c-47aa-947b-7865b045bf94)\"" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" Feb 02 15:47:48 crc kubenswrapper[4733]: I0202 15:47:48.255574 4733 scope.go:117] "RemoveContainer" containerID="8837c504005a41d8189bc30082d6c198635f60c659df3e0066907bf203ae6b97" Feb 02 15:47:48 crc kubenswrapper[4733]: E0202 15:47:48.256394 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zgsv9_openshift-machine-config-operator(83d0e810-ea5c-47aa-947b-7865b045bf94)\"" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" Feb 02 15:48:01 crc kubenswrapper[4733]: I0202 15:48:01.256333 4733 scope.go:117] "RemoveContainer" containerID="8837c504005a41d8189bc30082d6c198635f60c659df3e0066907bf203ae6b97" Feb 02 15:48:01 crc kubenswrapper[4733]: E0202 15:48:01.257272 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zgsv9_openshift-machine-config-operator(83d0e810-ea5c-47aa-947b-7865b045bf94)\"" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" Feb 02 15:48:15 crc kubenswrapper[4733]: I0202 15:48:15.255806 4733 scope.go:117] "RemoveContainer" containerID="8837c504005a41d8189bc30082d6c198635f60c659df3e0066907bf203ae6b97" Feb 02 15:48:15 crc kubenswrapper[4733]: E0202 15:48:15.256851 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zgsv9_openshift-machine-config-operator(83d0e810-ea5c-47aa-947b-7865b045bf94)\"" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" Feb 02 15:48:17 crc kubenswrapper[4733]: I0202 15:48:17.039140 4733 generic.go:334] "Generic (PLEG): container finished" podID="fdb2080f-f5b6-4f5f-9aeb-44b4efaebca6" containerID="db9bfa7e886378a285fdbbd3809ecfbed7b05ccb83c467fe287b93437e0af37b" exitCode=0 Feb 02 15:48:17 crc kubenswrapper[4733]: I0202 15:48:17.039215 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jbls7/must-gather-jwwdr" event={"ID":"fdb2080f-f5b6-4f5f-9aeb-44b4efaebca6","Type":"ContainerDied","Data":"db9bfa7e886378a285fdbbd3809ecfbed7b05ccb83c467fe287b93437e0af37b"} Feb 02 15:48:17 crc kubenswrapper[4733]: I0202 15:48:17.039721 4733 scope.go:117] "RemoveContainer" containerID="db9bfa7e886378a285fdbbd3809ecfbed7b05ccb83c467fe287b93437e0af37b" Feb 02 15:48:17 crc kubenswrapper[4733]: I0202 15:48:17.874740 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-jbls7_must-gather-jwwdr_fdb2080f-f5b6-4f5f-9aeb-44b4efaebca6/gather/0.log" Feb 02 15:48:26 crc kubenswrapper[4733]: I0202 15:48:26.255760 4733 scope.go:117] "RemoveContainer" containerID="8837c504005a41d8189bc30082d6c198635f60c659df3e0066907bf203ae6b97" Feb 02 15:48:26 crc kubenswrapper[4733]: E0202 15:48:26.257236 4733 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zgsv9_openshift-machine-config-operator(83d0e810-ea5c-47aa-947b-7865b045bf94)\"" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" Feb 02 15:48:26 crc kubenswrapper[4733]: I0202 15:48:26.707691 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-jbls7/must-gather-jwwdr"] Feb 02 15:48:26 crc kubenswrapper[4733]: I0202 15:48:26.708204 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-jbls7/must-gather-jwwdr" podUID="fdb2080f-f5b6-4f5f-9aeb-44b4efaebca6" containerName="copy" containerID="cri-o://4e232227b19c2c7fd67d5324582dc813695db0ce368183f2a72fd2991c870915" gracePeriod=2 Feb 02 15:48:26 crc kubenswrapper[4733]: I0202 15:48:26.715884 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-jbls7/must-gather-jwwdr"] Feb 02 15:48:27 crc kubenswrapper[4733]: I0202 15:48:27.042117 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-jbls7_must-gather-jwwdr_fdb2080f-f5b6-4f5f-9aeb-44b4efaebca6/copy/0.log" Feb 02 15:48:27 crc kubenswrapper[4733]: I0202 15:48:27.042971 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jbls7/must-gather-jwwdr" Feb 02 15:48:27 crc kubenswrapper[4733]: I0202 15:48:27.078413 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/fdb2080f-f5b6-4f5f-9aeb-44b4efaebca6-must-gather-output\") pod \"fdb2080f-f5b6-4f5f-9aeb-44b4efaebca6\" (UID: \"fdb2080f-f5b6-4f5f-9aeb-44b4efaebca6\") " Feb 02 15:48:27 crc kubenswrapper[4733]: I0202 15:48:27.078484 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gclnp\" (UniqueName: \"kubernetes.io/projected/fdb2080f-f5b6-4f5f-9aeb-44b4efaebca6-kube-api-access-gclnp\") pod \"fdb2080f-f5b6-4f5f-9aeb-44b4efaebca6\" (UID: \"fdb2080f-f5b6-4f5f-9aeb-44b4efaebca6\") " Feb 02 15:48:27 crc kubenswrapper[4733]: I0202 15:48:27.084129 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fdb2080f-f5b6-4f5f-9aeb-44b4efaebca6-kube-api-access-gclnp" (OuterVolumeSpecName: "kube-api-access-gclnp") pod "fdb2080f-f5b6-4f5f-9aeb-44b4efaebca6" (UID: "fdb2080f-f5b6-4f5f-9aeb-44b4efaebca6"). InnerVolumeSpecName "kube-api-access-gclnp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:48:27 crc kubenswrapper[4733]: I0202 15:48:27.117604 4733 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-jbls7_must-gather-jwwdr_fdb2080f-f5b6-4f5f-9aeb-44b4efaebca6/copy/0.log" Feb 02 15:48:27 crc kubenswrapper[4733]: I0202 15:48:27.118132 4733 generic.go:334] "Generic (PLEG): container finished" podID="fdb2080f-f5b6-4f5f-9aeb-44b4efaebca6" containerID="4e232227b19c2c7fd67d5324582dc813695db0ce368183f2a72fd2991c870915" exitCode=143 Feb 02 15:48:27 crc kubenswrapper[4733]: I0202 15:48:27.118205 4733 scope.go:117] "RemoveContainer" containerID="4e232227b19c2c7fd67d5324582dc813695db0ce368183f2a72fd2991c870915" Feb 02 15:48:27 crc kubenswrapper[4733]: I0202 15:48:27.118333 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jbls7/must-gather-jwwdr" Feb 02 15:48:27 crc kubenswrapper[4733]: I0202 15:48:27.138596 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fdb2080f-f5b6-4f5f-9aeb-44b4efaebca6-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "fdb2080f-f5b6-4f5f-9aeb-44b4efaebca6" (UID: "fdb2080f-f5b6-4f5f-9aeb-44b4efaebca6"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:48:27 crc kubenswrapper[4733]: I0202 15:48:27.141687 4733 scope.go:117] "RemoveContainer" containerID="db9bfa7e886378a285fdbbd3809ecfbed7b05ccb83c467fe287b93437e0af37b" Feb 02 15:48:27 crc kubenswrapper[4733]: I0202 15:48:27.179886 4733 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/fdb2080f-f5b6-4f5f-9aeb-44b4efaebca6-must-gather-output\") on node \"crc\" DevicePath \"\"" Feb 02 15:48:27 crc kubenswrapper[4733]: I0202 15:48:27.179942 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gclnp\" (UniqueName: \"kubernetes.io/projected/fdb2080f-f5b6-4f5f-9aeb-44b4efaebca6-kube-api-access-gclnp\") on node \"crc\" DevicePath \"\"" Feb 02 15:48:27 crc kubenswrapper[4733]: I0202 15:48:27.196570 4733 scope.go:117] "RemoveContainer" containerID="4e232227b19c2c7fd67d5324582dc813695db0ce368183f2a72fd2991c870915" Feb 02 15:48:27 crc kubenswrapper[4733]: E0202 15:48:27.197032 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e232227b19c2c7fd67d5324582dc813695db0ce368183f2a72fd2991c870915\": container with ID starting with 4e232227b19c2c7fd67d5324582dc813695db0ce368183f2a72fd2991c870915 not found: ID does not exist" containerID="4e232227b19c2c7fd67d5324582dc813695db0ce368183f2a72fd2991c870915" Feb 02 15:48:27 crc kubenswrapper[4733]: I0202 15:48:27.197101 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e232227b19c2c7fd67d5324582dc813695db0ce368183f2a72fd2991c870915"} err="failed to get container status \"4e232227b19c2c7fd67d5324582dc813695db0ce368183f2a72fd2991c870915\": rpc error: code = NotFound desc = could not find container \"4e232227b19c2c7fd67d5324582dc813695db0ce368183f2a72fd2991c870915\": container with ID starting with 4e232227b19c2c7fd67d5324582dc813695db0ce368183f2a72fd2991c870915 not found: ID does not exist" Feb 02 15:48:27 crc kubenswrapper[4733]: I0202 15:48:27.197135 4733 scope.go:117] "RemoveContainer" containerID="db9bfa7e886378a285fdbbd3809ecfbed7b05ccb83c467fe287b93437e0af37b" Feb 02 15:48:27 crc kubenswrapper[4733]: E0202 15:48:27.197655 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db9bfa7e886378a285fdbbd3809ecfbed7b05ccb83c467fe287b93437e0af37b\": container with ID starting with db9bfa7e886378a285fdbbd3809ecfbed7b05ccb83c467fe287b93437e0af37b not found: ID does not exist" containerID="db9bfa7e886378a285fdbbd3809ecfbed7b05ccb83c467fe287b93437e0af37b" Feb 02 15:48:27 crc kubenswrapper[4733]: I0202 15:48:27.197723 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db9bfa7e886378a285fdbbd3809ecfbed7b05ccb83c467fe287b93437e0af37b"} err="failed to get container status \"db9bfa7e886378a285fdbbd3809ecfbed7b05ccb83c467fe287b93437e0af37b\": rpc error: code = NotFound desc = could not find container \"db9bfa7e886378a285fdbbd3809ecfbed7b05ccb83c467fe287b93437e0af37b\": container with ID starting with db9bfa7e886378a285fdbbd3809ecfbed7b05ccb83c467fe287b93437e0af37b not found: ID does not exist" Feb 02 15:48:27 crc kubenswrapper[4733]: I0202 15:48:27.267752 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fdb2080f-f5b6-4f5f-9aeb-44b4efaebca6" path="/var/lib/kubelet/pods/fdb2080f-f5b6-4f5f-9aeb-44b4efaebca6/volumes" Feb 02 15:48:31 crc kubenswrapper[4733]: I0202 15:48:31.332738 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qb54w"] Feb 02 15:48:31 crc kubenswrapper[4733]: E0202 15:48:31.333646 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdb2080f-f5b6-4f5f-9aeb-44b4efaebca6" containerName="gather" Feb 02 15:48:31 crc kubenswrapper[4733]: I0202 15:48:31.333665 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdb2080f-f5b6-4f5f-9aeb-44b4efaebca6" containerName="gather" Feb 02 15:48:31 crc kubenswrapper[4733]: E0202 15:48:31.333682 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdb2080f-f5b6-4f5f-9aeb-44b4efaebca6" containerName="copy" Feb 02 15:48:31 crc kubenswrapper[4733]: I0202 15:48:31.333688 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdb2080f-f5b6-4f5f-9aeb-44b4efaebca6" containerName="copy" Feb 02 15:48:31 crc kubenswrapper[4733]: I0202 15:48:31.333838 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdb2080f-f5b6-4f5f-9aeb-44b4efaebca6" containerName="copy" Feb 02 15:48:31 crc kubenswrapper[4733]: I0202 15:48:31.333851 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdb2080f-f5b6-4f5f-9aeb-44b4efaebca6" containerName="gather" Feb 02 15:48:31 crc kubenswrapper[4733]: I0202 15:48:31.334850 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qb54w" Feb 02 15:48:31 crc kubenswrapper[4733]: I0202 15:48:31.349074 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qb54w"] Feb 02 15:48:31 crc kubenswrapper[4733]: I0202 15:48:31.437634 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfbvm\" (UniqueName: \"kubernetes.io/projected/00cd0143-f4ee-4d0b-88f7-b44e94782ae7-kube-api-access-xfbvm\") pod \"redhat-operators-qb54w\" (UID: \"00cd0143-f4ee-4d0b-88f7-b44e94782ae7\") " pod="openshift-marketplace/redhat-operators-qb54w" Feb 02 15:48:31 crc kubenswrapper[4733]: I0202 15:48:31.437814 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00cd0143-f4ee-4d0b-88f7-b44e94782ae7-catalog-content\") pod \"redhat-operators-qb54w\" (UID: \"00cd0143-f4ee-4d0b-88f7-b44e94782ae7\") " pod="openshift-marketplace/redhat-operators-qb54w" Feb 02 15:48:31 crc kubenswrapper[4733]: I0202 15:48:31.437887 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00cd0143-f4ee-4d0b-88f7-b44e94782ae7-utilities\") pod \"redhat-operators-qb54w\" (UID: \"00cd0143-f4ee-4d0b-88f7-b44e94782ae7\") " pod="openshift-marketplace/redhat-operators-qb54w" Feb 02 15:48:31 crc kubenswrapper[4733]: I0202 15:48:31.539360 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfbvm\" (UniqueName: \"kubernetes.io/projected/00cd0143-f4ee-4d0b-88f7-b44e94782ae7-kube-api-access-xfbvm\") pod \"redhat-operators-qb54w\" (UID: \"00cd0143-f4ee-4d0b-88f7-b44e94782ae7\") " pod="openshift-marketplace/redhat-operators-qb54w" Feb 02 15:48:31 crc kubenswrapper[4733]: I0202 15:48:31.539422 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00cd0143-f4ee-4d0b-88f7-b44e94782ae7-catalog-content\") pod \"redhat-operators-qb54w\" (UID: \"00cd0143-f4ee-4d0b-88f7-b44e94782ae7\") " pod="openshift-marketplace/redhat-operators-qb54w" Feb 02 15:48:31 crc kubenswrapper[4733]: I0202 15:48:31.539444 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00cd0143-f4ee-4d0b-88f7-b44e94782ae7-utilities\") pod \"redhat-operators-qb54w\" (UID: \"00cd0143-f4ee-4d0b-88f7-b44e94782ae7\") " pod="openshift-marketplace/redhat-operators-qb54w" Feb 02 15:48:31 crc kubenswrapper[4733]: I0202 15:48:31.539902 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00cd0143-f4ee-4d0b-88f7-b44e94782ae7-utilities\") pod \"redhat-operators-qb54w\" (UID: \"00cd0143-f4ee-4d0b-88f7-b44e94782ae7\") " pod="openshift-marketplace/redhat-operators-qb54w" Feb 02 15:48:31 crc kubenswrapper[4733]: I0202 15:48:31.539922 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00cd0143-f4ee-4d0b-88f7-b44e94782ae7-catalog-content\") pod \"redhat-operators-qb54w\" (UID: \"00cd0143-f4ee-4d0b-88f7-b44e94782ae7\") " pod="openshift-marketplace/redhat-operators-qb54w" Feb 02 15:48:31 crc kubenswrapper[4733]: I0202 15:48:31.565691 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfbvm\" (UniqueName: \"kubernetes.io/projected/00cd0143-f4ee-4d0b-88f7-b44e94782ae7-kube-api-access-xfbvm\") pod \"redhat-operators-qb54w\" (UID: \"00cd0143-f4ee-4d0b-88f7-b44e94782ae7\") " pod="openshift-marketplace/redhat-operators-qb54w" Feb 02 15:48:31 crc kubenswrapper[4733]: I0202 15:48:31.682431 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qb54w" Feb 02 15:48:32 crc kubenswrapper[4733]: I0202 15:48:32.081141 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qb54w"] Feb 02 15:48:32 crc kubenswrapper[4733]: I0202 15:48:32.157825 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qb54w" event={"ID":"00cd0143-f4ee-4d0b-88f7-b44e94782ae7","Type":"ContainerStarted","Data":"547f9a289fce1dcddc9fd0a97e00e417968e7ecec170fcb16127b329fdd45c56"} Feb 02 15:48:33 crc kubenswrapper[4733]: I0202 15:48:33.168492 4733 generic.go:334] "Generic (PLEG): container finished" podID="00cd0143-f4ee-4d0b-88f7-b44e94782ae7" containerID="cfb7f5e4eb43126ad6b58894f4213905826acb300fe30e3d8638a4995dd80166" exitCode=0 Feb 02 15:48:33 crc kubenswrapper[4733]: I0202 15:48:33.168583 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qb54w" event={"ID":"00cd0143-f4ee-4d0b-88f7-b44e94782ae7","Type":"ContainerDied","Data":"cfb7f5e4eb43126ad6b58894f4213905826acb300fe30e3d8638a4995dd80166"} Feb 02 15:48:34 crc kubenswrapper[4733]: I0202 15:48:34.180311 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qb54w" event={"ID":"00cd0143-f4ee-4d0b-88f7-b44e94782ae7","Type":"ContainerStarted","Data":"8bf6e70dd7784fdb57376feed68be9de7b3ff681f4abecef2b6ec70c5533dc60"} Feb 02 15:48:35 crc kubenswrapper[4733]: I0202 15:48:35.190442 4733 generic.go:334] "Generic (PLEG): container finished" podID="00cd0143-f4ee-4d0b-88f7-b44e94782ae7" containerID="8bf6e70dd7784fdb57376feed68be9de7b3ff681f4abecef2b6ec70c5533dc60" exitCode=0 Feb 02 15:48:35 crc kubenswrapper[4733]: I0202 15:48:35.190525 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qb54w" event={"ID":"00cd0143-f4ee-4d0b-88f7-b44e94782ae7","Type":"ContainerDied","Data":"8bf6e70dd7784fdb57376feed68be9de7b3ff681f4abecef2b6ec70c5533dc60"} Feb 02 15:48:36 crc kubenswrapper[4733]: I0202 15:48:36.201603 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qb54w" event={"ID":"00cd0143-f4ee-4d0b-88f7-b44e94782ae7","Type":"ContainerStarted","Data":"c11d0194eac3d67d7502b7176f55a59e40713246843f811c2b503236499f019e"} Feb 02 15:48:36 crc kubenswrapper[4733]: I0202 15:48:36.240360 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qb54w" podStartSLOduration=2.725937207 podStartE2EDuration="5.240322287s" podCreationTimestamp="2026-02-02 15:48:31 +0000 UTC" firstStartedPulling="2026-02-02 15:48:33.17065349 +0000 UTC m=+1996.622114858" lastFinishedPulling="2026-02-02 15:48:35.68503855 +0000 UTC m=+1999.136499938" observedRunningTime="2026-02-02 15:48:36.229644756 +0000 UTC m=+1999.681106144" watchObservedRunningTime="2026-02-02 15:48:36.240322287 +0000 UTC m=+1999.691783685" Feb 02 15:48:38 crc kubenswrapper[4733]: I0202 15:48:38.257752 4733 scope.go:117] "RemoveContainer" containerID="8837c504005a41d8189bc30082d6c198635f60c659df3e0066907bf203ae6b97" Feb 02 15:48:39 crc kubenswrapper[4733]: I0202 15:48:39.219638 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" event={"ID":"83d0e810-ea5c-47aa-947b-7865b045bf94","Type":"ContainerStarted","Data":"574d37318bd4cb5a98a95edbec113cd09aff4a55e83b87692c0ee728d199b6bd"} Feb 02 15:48:41 crc kubenswrapper[4733]: I0202 15:48:41.683030 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qb54w" Feb 02 15:48:41 crc kubenswrapper[4733]: I0202 15:48:41.683348 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qb54w" Feb 02 15:48:42 crc kubenswrapper[4733]: I0202 15:48:42.739944 4733 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qb54w" podUID="00cd0143-f4ee-4d0b-88f7-b44e94782ae7" containerName="registry-server" probeResult="failure" output=< Feb 02 15:48:42 crc kubenswrapper[4733]: timeout: failed to connect service ":50051" within 1s Feb 02 15:48:42 crc kubenswrapper[4733]: > Feb 02 15:48:51 crc kubenswrapper[4733]: I0202 15:48:51.726214 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qb54w" Feb 02 15:48:51 crc kubenswrapper[4733]: I0202 15:48:51.784040 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qb54w" Feb 02 15:48:51 crc kubenswrapper[4733]: I0202 15:48:51.977742 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qb54w"] Feb 02 15:48:53 crc kubenswrapper[4733]: I0202 15:48:53.320100 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qb54w" podUID="00cd0143-f4ee-4d0b-88f7-b44e94782ae7" containerName="registry-server" containerID="cri-o://c11d0194eac3d67d7502b7176f55a59e40713246843f811c2b503236499f019e" gracePeriod=2 Feb 02 15:48:53 crc kubenswrapper[4733]: I0202 15:48:53.701432 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qb54w" Feb 02 15:48:53 crc kubenswrapper[4733]: I0202 15:48:53.745808 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xfbvm\" (UniqueName: \"kubernetes.io/projected/00cd0143-f4ee-4d0b-88f7-b44e94782ae7-kube-api-access-xfbvm\") pod \"00cd0143-f4ee-4d0b-88f7-b44e94782ae7\" (UID: \"00cd0143-f4ee-4d0b-88f7-b44e94782ae7\") " Feb 02 15:48:53 crc kubenswrapper[4733]: I0202 15:48:53.745869 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00cd0143-f4ee-4d0b-88f7-b44e94782ae7-catalog-content\") pod \"00cd0143-f4ee-4d0b-88f7-b44e94782ae7\" (UID: \"00cd0143-f4ee-4d0b-88f7-b44e94782ae7\") " Feb 02 15:48:53 crc kubenswrapper[4733]: I0202 15:48:53.745931 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00cd0143-f4ee-4d0b-88f7-b44e94782ae7-utilities\") pod \"00cd0143-f4ee-4d0b-88f7-b44e94782ae7\" (UID: \"00cd0143-f4ee-4d0b-88f7-b44e94782ae7\") " Feb 02 15:48:53 crc kubenswrapper[4733]: I0202 15:48:53.746782 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00cd0143-f4ee-4d0b-88f7-b44e94782ae7-utilities" (OuterVolumeSpecName: "utilities") pod "00cd0143-f4ee-4d0b-88f7-b44e94782ae7" (UID: "00cd0143-f4ee-4d0b-88f7-b44e94782ae7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:48:53 crc kubenswrapper[4733]: I0202 15:48:53.754008 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00cd0143-f4ee-4d0b-88f7-b44e94782ae7-kube-api-access-xfbvm" (OuterVolumeSpecName: "kube-api-access-xfbvm") pod "00cd0143-f4ee-4d0b-88f7-b44e94782ae7" (UID: "00cd0143-f4ee-4d0b-88f7-b44e94782ae7"). InnerVolumeSpecName "kube-api-access-xfbvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:48:53 crc kubenswrapper[4733]: I0202 15:48:53.847660 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xfbvm\" (UniqueName: \"kubernetes.io/projected/00cd0143-f4ee-4d0b-88f7-b44e94782ae7-kube-api-access-xfbvm\") on node \"crc\" DevicePath \"\"" Feb 02 15:48:53 crc kubenswrapper[4733]: I0202 15:48:53.847698 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00cd0143-f4ee-4d0b-88f7-b44e94782ae7-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 15:48:53 crc kubenswrapper[4733]: I0202 15:48:53.909347 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00cd0143-f4ee-4d0b-88f7-b44e94782ae7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "00cd0143-f4ee-4d0b-88f7-b44e94782ae7" (UID: "00cd0143-f4ee-4d0b-88f7-b44e94782ae7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:48:53 crc kubenswrapper[4733]: I0202 15:48:53.948534 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00cd0143-f4ee-4d0b-88f7-b44e94782ae7-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 15:48:54 crc kubenswrapper[4733]: I0202 15:48:54.331510 4733 generic.go:334] "Generic (PLEG): container finished" podID="00cd0143-f4ee-4d0b-88f7-b44e94782ae7" containerID="c11d0194eac3d67d7502b7176f55a59e40713246843f811c2b503236499f019e" exitCode=0 Feb 02 15:48:54 crc kubenswrapper[4733]: I0202 15:48:54.331561 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qb54w" event={"ID":"00cd0143-f4ee-4d0b-88f7-b44e94782ae7","Type":"ContainerDied","Data":"c11d0194eac3d67d7502b7176f55a59e40713246843f811c2b503236499f019e"} Feb 02 15:48:54 crc kubenswrapper[4733]: I0202 15:48:54.331592 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qb54w" event={"ID":"00cd0143-f4ee-4d0b-88f7-b44e94782ae7","Type":"ContainerDied","Data":"547f9a289fce1dcddc9fd0a97e00e417968e7ecec170fcb16127b329fdd45c56"} Feb 02 15:48:54 crc kubenswrapper[4733]: I0202 15:48:54.331616 4733 scope.go:117] "RemoveContainer" containerID="c11d0194eac3d67d7502b7176f55a59e40713246843f811c2b503236499f019e" Feb 02 15:48:54 crc kubenswrapper[4733]: I0202 15:48:54.331609 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qb54w" Feb 02 15:48:54 crc kubenswrapper[4733]: I0202 15:48:54.354309 4733 scope.go:117] "RemoveContainer" containerID="8bf6e70dd7784fdb57376feed68be9de7b3ff681f4abecef2b6ec70c5533dc60" Feb 02 15:48:54 crc kubenswrapper[4733]: I0202 15:48:54.377223 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qb54w"] Feb 02 15:48:54 crc kubenswrapper[4733]: I0202 15:48:54.382898 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qb54w"] Feb 02 15:48:54 crc kubenswrapper[4733]: I0202 15:48:54.391575 4733 scope.go:117] "RemoveContainer" containerID="cfb7f5e4eb43126ad6b58894f4213905826acb300fe30e3d8638a4995dd80166" Feb 02 15:48:54 crc kubenswrapper[4733]: I0202 15:48:54.412699 4733 scope.go:117] "RemoveContainer" containerID="c11d0194eac3d67d7502b7176f55a59e40713246843f811c2b503236499f019e" Feb 02 15:48:54 crc kubenswrapper[4733]: E0202 15:48:54.413061 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c11d0194eac3d67d7502b7176f55a59e40713246843f811c2b503236499f019e\": container with ID starting with c11d0194eac3d67d7502b7176f55a59e40713246843f811c2b503236499f019e not found: ID does not exist" containerID="c11d0194eac3d67d7502b7176f55a59e40713246843f811c2b503236499f019e" Feb 02 15:48:54 crc kubenswrapper[4733]: I0202 15:48:54.413090 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c11d0194eac3d67d7502b7176f55a59e40713246843f811c2b503236499f019e"} err="failed to get container status \"c11d0194eac3d67d7502b7176f55a59e40713246843f811c2b503236499f019e\": rpc error: code = NotFound desc = could not find container \"c11d0194eac3d67d7502b7176f55a59e40713246843f811c2b503236499f019e\": container with ID starting with c11d0194eac3d67d7502b7176f55a59e40713246843f811c2b503236499f019e not found: ID does not exist" Feb 02 15:48:54 crc kubenswrapper[4733]: I0202 15:48:54.413109 4733 scope.go:117] "RemoveContainer" containerID="8bf6e70dd7784fdb57376feed68be9de7b3ff681f4abecef2b6ec70c5533dc60" Feb 02 15:48:54 crc kubenswrapper[4733]: E0202 15:48:54.413598 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8bf6e70dd7784fdb57376feed68be9de7b3ff681f4abecef2b6ec70c5533dc60\": container with ID starting with 8bf6e70dd7784fdb57376feed68be9de7b3ff681f4abecef2b6ec70c5533dc60 not found: ID does not exist" containerID="8bf6e70dd7784fdb57376feed68be9de7b3ff681f4abecef2b6ec70c5533dc60" Feb 02 15:48:54 crc kubenswrapper[4733]: I0202 15:48:54.413816 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8bf6e70dd7784fdb57376feed68be9de7b3ff681f4abecef2b6ec70c5533dc60"} err="failed to get container status \"8bf6e70dd7784fdb57376feed68be9de7b3ff681f4abecef2b6ec70c5533dc60\": rpc error: code = NotFound desc = could not find container \"8bf6e70dd7784fdb57376feed68be9de7b3ff681f4abecef2b6ec70c5533dc60\": container with ID starting with 8bf6e70dd7784fdb57376feed68be9de7b3ff681f4abecef2b6ec70c5533dc60 not found: ID does not exist" Feb 02 15:48:54 crc kubenswrapper[4733]: I0202 15:48:54.413978 4733 scope.go:117] "RemoveContainer" containerID="cfb7f5e4eb43126ad6b58894f4213905826acb300fe30e3d8638a4995dd80166" Feb 02 15:48:54 crc kubenswrapper[4733]: E0202 15:48:54.414517 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cfb7f5e4eb43126ad6b58894f4213905826acb300fe30e3d8638a4995dd80166\": container with ID starting with cfb7f5e4eb43126ad6b58894f4213905826acb300fe30e3d8638a4995dd80166 not found: ID does not exist" containerID="cfb7f5e4eb43126ad6b58894f4213905826acb300fe30e3d8638a4995dd80166" Feb 02 15:48:54 crc kubenswrapper[4733]: I0202 15:48:54.414552 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfb7f5e4eb43126ad6b58894f4213905826acb300fe30e3d8638a4995dd80166"} err="failed to get container status \"cfb7f5e4eb43126ad6b58894f4213905826acb300fe30e3d8638a4995dd80166\": rpc error: code = NotFound desc = could not find container \"cfb7f5e4eb43126ad6b58894f4213905826acb300fe30e3d8638a4995dd80166\": container with ID starting with cfb7f5e4eb43126ad6b58894f4213905826acb300fe30e3d8638a4995dd80166 not found: ID does not exist" Feb 02 15:48:55 crc kubenswrapper[4733]: I0202 15:48:55.270486 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00cd0143-f4ee-4d0b-88f7-b44e94782ae7" path="/var/lib/kubelet/pods/00cd0143-f4ee-4d0b-88f7-b44e94782ae7/volumes" Feb 02 15:49:58 crc kubenswrapper[4733]: I0202 15:49:58.174337 4733 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-sfv6f"] Feb 02 15:49:58 crc kubenswrapper[4733]: E0202 15:49:58.175760 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00cd0143-f4ee-4d0b-88f7-b44e94782ae7" containerName="registry-server" Feb 02 15:49:58 crc kubenswrapper[4733]: I0202 15:49:58.175825 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="00cd0143-f4ee-4d0b-88f7-b44e94782ae7" containerName="registry-server" Feb 02 15:49:58 crc kubenswrapper[4733]: E0202 15:49:58.175947 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00cd0143-f4ee-4d0b-88f7-b44e94782ae7" containerName="extract-content" Feb 02 15:49:58 crc kubenswrapper[4733]: I0202 15:49:58.176050 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="00cd0143-f4ee-4d0b-88f7-b44e94782ae7" containerName="extract-content" Feb 02 15:49:58 crc kubenswrapper[4733]: E0202 15:49:58.176080 4733 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00cd0143-f4ee-4d0b-88f7-b44e94782ae7" containerName="extract-utilities" Feb 02 15:49:58 crc kubenswrapper[4733]: I0202 15:49:58.176534 4733 state_mem.go:107] "Deleted CPUSet assignment" podUID="00cd0143-f4ee-4d0b-88f7-b44e94782ae7" containerName="extract-utilities" Feb 02 15:49:58 crc kubenswrapper[4733]: I0202 15:49:58.176953 4733 memory_manager.go:354] "RemoveStaleState removing state" podUID="00cd0143-f4ee-4d0b-88f7-b44e94782ae7" containerName="registry-server" Feb 02 15:49:58 crc kubenswrapper[4733]: I0202 15:49:58.180185 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sfv6f" Feb 02 15:49:58 crc kubenswrapper[4733]: I0202 15:49:58.191985 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sfv6f"] Feb 02 15:49:58 crc kubenswrapper[4733]: I0202 15:49:58.273991 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6498c564-0dc9-4c8b-aadf-a8d921c560c5-utilities\") pod \"community-operators-sfv6f\" (UID: \"6498c564-0dc9-4c8b-aadf-a8d921c560c5\") " pod="openshift-marketplace/community-operators-sfv6f" Feb 02 15:49:58 crc kubenswrapper[4733]: I0202 15:49:58.274444 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4mhj\" (UniqueName: \"kubernetes.io/projected/6498c564-0dc9-4c8b-aadf-a8d921c560c5-kube-api-access-p4mhj\") pod \"community-operators-sfv6f\" (UID: \"6498c564-0dc9-4c8b-aadf-a8d921c560c5\") " pod="openshift-marketplace/community-operators-sfv6f" Feb 02 15:49:58 crc kubenswrapper[4733]: I0202 15:49:58.274518 4733 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6498c564-0dc9-4c8b-aadf-a8d921c560c5-catalog-content\") pod \"community-operators-sfv6f\" (UID: \"6498c564-0dc9-4c8b-aadf-a8d921c560c5\") " pod="openshift-marketplace/community-operators-sfv6f" Feb 02 15:49:58 crc kubenswrapper[4733]: I0202 15:49:58.375103 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6498c564-0dc9-4c8b-aadf-a8d921c560c5-utilities\") pod \"community-operators-sfv6f\" (UID: \"6498c564-0dc9-4c8b-aadf-a8d921c560c5\") " pod="openshift-marketplace/community-operators-sfv6f" Feb 02 15:49:58 crc kubenswrapper[4733]: I0202 15:49:58.375210 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4mhj\" (UniqueName: \"kubernetes.io/projected/6498c564-0dc9-4c8b-aadf-a8d921c560c5-kube-api-access-p4mhj\") pod \"community-operators-sfv6f\" (UID: \"6498c564-0dc9-4c8b-aadf-a8d921c560c5\") " pod="openshift-marketplace/community-operators-sfv6f" Feb 02 15:49:58 crc kubenswrapper[4733]: I0202 15:49:58.375257 4733 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6498c564-0dc9-4c8b-aadf-a8d921c560c5-catalog-content\") pod \"community-operators-sfv6f\" (UID: \"6498c564-0dc9-4c8b-aadf-a8d921c560c5\") " pod="openshift-marketplace/community-operators-sfv6f" Feb 02 15:49:58 crc kubenswrapper[4733]: I0202 15:49:58.375648 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6498c564-0dc9-4c8b-aadf-a8d921c560c5-utilities\") pod \"community-operators-sfv6f\" (UID: \"6498c564-0dc9-4c8b-aadf-a8d921c560c5\") " pod="openshift-marketplace/community-operators-sfv6f" Feb 02 15:49:58 crc kubenswrapper[4733]: I0202 15:49:58.375781 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6498c564-0dc9-4c8b-aadf-a8d921c560c5-catalog-content\") pod \"community-operators-sfv6f\" (UID: \"6498c564-0dc9-4c8b-aadf-a8d921c560c5\") " pod="openshift-marketplace/community-operators-sfv6f" Feb 02 15:49:58 crc kubenswrapper[4733]: I0202 15:49:58.397732 4733 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4mhj\" (UniqueName: \"kubernetes.io/projected/6498c564-0dc9-4c8b-aadf-a8d921c560c5-kube-api-access-p4mhj\") pod \"community-operators-sfv6f\" (UID: \"6498c564-0dc9-4c8b-aadf-a8d921c560c5\") " pod="openshift-marketplace/community-operators-sfv6f" Feb 02 15:49:58 crc kubenswrapper[4733]: I0202 15:49:58.512706 4733 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sfv6f" Feb 02 15:49:59 crc kubenswrapper[4733]: I0202 15:49:59.019661 4733 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sfv6f"] Feb 02 15:49:59 crc kubenswrapper[4733]: I0202 15:49:59.850075 4733 generic.go:334] "Generic (PLEG): container finished" podID="6498c564-0dc9-4c8b-aadf-a8d921c560c5" containerID="3ac56d8ed4e4b3a07a44435c07e55954da4a7c4eb30f4cf55fd053881dd69a64" exitCode=0 Feb 02 15:49:59 crc kubenswrapper[4733]: I0202 15:49:59.850347 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sfv6f" event={"ID":"6498c564-0dc9-4c8b-aadf-a8d921c560c5","Type":"ContainerDied","Data":"3ac56d8ed4e4b3a07a44435c07e55954da4a7c4eb30f4cf55fd053881dd69a64"} Feb 02 15:49:59 crc kubenswrapper[4733]: I0202 15:49:59.850376 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sfv6f" event={"ID":"6498c564-0dc9-4c8b-aadf-a8d921c560c5","Type":"ContainerStarted","Data":"7ac873a7b850ab508c80be570a3cd93174c4fe98cdad91a8fec6a5227a5fa1a2"} Feb 02 15:49:59 crc kubenswrapper[4733]: I0202 15:49:59.852071 4733 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 02 15:50:00 crc kubenswrapper[4733]: I0202 15:50:00.859883 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sfv6f" event={"ID":"6498c564-0dc9-4c8b-aadf-a8d921c560c5","Type":"ContainerStarted","Data":"fac49fae91c304d5ba93f5ac1352af229c9ae06429860065c7979d5bec4203c6"} Feb 02 15:50:01 crc kubenswrapper[4733]: I0202 15:50:01.869994 4733 generic.go:334] "Generic (PLEG): container finished" podID="6498c564-0dc9-4c8b-aadf-a8d921c560c5" containerID="fac49fae91c304d5ba93f5ac1352af229c9ae06429860065c7979d5bec4203c6" exitCode=0 Feb 02 15:50:01 crc kubenswrapper[4733]: I0202 15:50:01.870044 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sfv6f" event={"ID":"6498c564-0dc9-4c8b-aadf-a8d921c560c5","Type":"ContainerDied","Data":"fac49fae91c304d5ba93f5ac1352af229c9ae06429860065c7979d5bec4203c6"} Feb 02 15:50:02 crc kubenswrapper[4733]: I0202 15:50:02.883813 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sfv6f" event={"ID":"6498c564-0dc9-4c8b-aadf-a8d921c560c5","Type":"ContainerStarted","Data":"884a7625c02fed00db769ed10f5c30e096f55ef9b572b71f5d3f38a42c7f5ecc"} Feb 02 15:50:02 crc kubenswrapper[4733]: I0202 15:50:02.923678 4733 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-sfv6f" podStartSLOduration=2.505082828 podStartE2EDuration="4.923646522s" podCreationTimestamp="2026-02-02 15:49:58 +0000 UTC" firstStartedPulling="2026-02-02 15:49:59.851805487 +0000 UTC m=+2083.303266855" lastFinishedPulling="2026-02-02 15:50:02.270369181 +0000 UTC m=+2085.721830549" observedRunningTime="2026-02-02 15:50:02.916828983 +0000 UTC m=+2086.368290351" watchObservedRunningTime="2026-02-02 15:50:02.923646522 +0000 UTC m=+2086.375107920" Feb 02 15:50:08 crc kubenswrapper[4733]: I0202 15:50:08.513599 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-sfv6f" Feb 02 15:50:08 crc kubenswrapper[4733]: I0202 15:50:08.513991 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-sfv6f" Feb 02 15:50:08 crc kubenswrapper[4733]: I0202 15:50:08.581566 4733 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-sfv6f" Feb 02 15:50:08 crc kubenswrapper[4733]: I0202 15:50:08.967898 4733 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-sfv6f" Feb 02 15:50:09 crc kubenswrapper[4733]: I0202 15:50:09.020670 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-sfv6f"] Feb 02 15:50:10 crc kubenswrapper[4733]: I0202 15:50:10.936565 4733 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-sfv6f" podUID="6498c564-0dc9-4c8b-aadf-a8d921c560c5" containerName="registry-server" containerID="cri-o://884a7625c02fed00db769ed10f5c30e096f55ef9b572b71f5d3f38a42c7f5ecc" gracePeriod=2 Feb 02 15:50:11 crc kubenswrapper[4733]: I0202 15:50:11.450544 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sfv6f" Feb 02 15:50:11 crc kubenswrapper[4733]: I0202 15:50:11.457888 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6498c564-0dc9-4c8b-aadf-a8d921c560c5-catalog-content\") pod \"6498c564-0dc9-4c8b-aadf-a8d921c560c5\" (UID: \"6498c564-0dc9-4c8b-aadf-a8d921c560c5\") " Feb 02 15:50:11 crc kubenswrapper[4733]: I0202 15:50:11.457993 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4mhj\" (UniqueName: \"kubernetes.io/projected/6498c564-0dc9-4c8b-aadf-a8d921c560c5-kube-api-access-p4mhj\") pod \"6498c564-0dc9-4c8b-aadf-a8d921c560c5\" (UID: \"6498c564-0dc9-4c8b-aadf-a8d921c560c5\") " Feb 02 15:50:11 crc kubenswrapper[4733]: I0202 15:50:11.458072 4733 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6498c564-0dc9-4c8b-aadf-a8d921c560c5-utilities\") pod \"6498c564-0dc9-4c8b-aadf-a8d921c560c5\" (UID: \"6498c564-0dc9-4c8b-aadf-a8d921c560c5\") " Feb 02 15:50:11 crc kubenswrapper[4733]: I0202 15:50:11.461073 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6498c564-0dc9-4c8b-aadf-a8d921c560c5-utilities" (OuterVolumeSpecName: "utilities") pod "6498c564-0dc9-4c8b-aadf-a8d921c560c5" (UID: "6498c564-0dc9-4c8b-aadf-a8d921c560c5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:50:11 crc kubenswrapper[4733]: I0202 15:50:11.468378 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6498c564-0dc9-4c8b-aadf-a8d921c560c5-kube-api-access-p4mhj" (OuterVolumeSpecName: "kube-api-access-p4mhj") pod "6498c564-0dc9-4c8b-aadf-a8d921c560c5" (UID: "6498c564-0dc9-4c8b-aadf-a8d921c560c5"). InnerVolumeSpecName "kube-api-access-p4mhj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 02 15:50:11 crc kubenswrapper[4733]: I0202 15:50:11.516857 4733 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6498c564-0dc9-4c8b-aadf-a8d921c560c5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6498c564-0dc9-4c8b-aadf-a8d921c560c5" (UID: "6498c564-0dc9-4c8b-aadf-a8d921c560c5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 02 15:50:11 crc kubenswrapper[4733]: I0202 15:50:11.560533 4733 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4mhj\" (UniqueName: \"kubernetes.io/projected/6498c564-0dc9-4c8b-aadf-a8d921c560c5-kube-api-access-p4mhj\") on node \"crc\" DevicePath \"\"" Feb 02 15:50:11 crc kubenswrapper[4733]: I0202 15:50:11.560592 4733 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6498c564-0dc9-4c8b-aadf-a8d921c560c5-utilities\") on node \"crc\" DevicePath \"\"" Feb 02 15:50:11 crc kubenswrapper[4733]: I0202 15:50:11.560613 4733 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6498c564-0dc9-4c8b-aadf-a8d921c560c5-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 02 15:50:11 crc kubenswrapper[4733]: I0202 15:50:11.947284 4733 generic.go:334] "Generic (PLEG): container finished" podID="6498c564-0dc9-4c8b-aadf-a8d921c560c5" containerID="884a7625c02fed00db769ed10f5c30e096f55ef9b572b71f5d3f38a42c7f5ecc" exitCode=0 Feb 02 15:50:11 crc kubenswrapper[4733]: I0202 15:50:11.947361 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sfv6f" event={"ID":"6498c564-0dc9-4c8b-aadf-a8d921c560c5","Type":"ContainerDied","Data":"884a7625c02fed00db769ed10f5c30e096f55ef9b572b71f5d3f38a42c7f5ecc"} Feb 02 15:50:11 crc kubenswrapper[4733]: I0202 15:50:11.947733 4733 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sfv6f" event={"ID":"6498c564-0dc9-4c8b-aadf-a8d921c560c5","Type":"ContainerDied","Data":"7ac873a7b850ab508c80be570a3cd93174c4fe98cdad91a8fec6a5227a5fa1a2"} Feb 02 15:50:11 crc kubenswrapper[4733]: I0202 15:50:11.947769 4733 scope.go:117] "RemoveContainer" containerID="884a7625c02fed00db769ed10f5c30e096f55ef9b572b71f5d3f38a42c7f5ecc" Feb 02 15:50:11 crc kubenswrapper[4733]: I0202 15:50:11.947383 4733 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sfv6f" Feb 02 15:50:11 crc kubenswrapper[4733]: I0202 15:50:11.971653 4733 scope.go:117] "RemoveContainer" containerID="fac49fae91c304d5ba93f5ac1352af229c9ae06429860065c7979d5bec4203c6" Feb 02 15:50:11 crc kubenswrapper[4733]: I0202 15:50:11.991381 4733 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-sfv6f"] Feb 02 15:50:11 crc kubenswrapper[4733]: I0202 15:50:11.996379 4733 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-sfv6f"] Feb 02 15:50:12 crc kubenswrapper[4733]: I0202 15:50:12.025942 4733 scope.go:117] "RemoveContainer" containerID="3ac56d8ed4e4b3a07a44435c07e55954da4a7c4eb30f4cf55fd053881dd69a64" Feb 02 15:50:12 crc kubenswrapper[4733]: I0202 15:50:12.045151 4733 scope.go:117] "RemoveContainer" containerID="884a7625c02fed00db769ed10f5c30e096f55ef9b572b71f5d3f38a42c7f5ecc" Feb 02 15:50:12 crc kubenswrapper[4733]: E0202 15:50:12.045656 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"884a7625c02fed00db769ed10f5c30e096f55ef9b572b71f5d3f38a42c7f5ecc\": container with ID starting with 884a7625c02fed00db769ed10f5c30e096f55ef9b572b71f5d3f38a42c7f5ecc not found: ID does not exist" containerID="884a7625c02fed00db769ed10f5c30e096f55ef9b572b71f5d3f38a42c7f5ecc" Feb 02 15:50:12 crc kubenswrapper[4733]: I0202 15:50:12.045803 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"884a7625c02fed00db769ed10f5c30e096f55ef9b572b71f5d3f38a42c7f5ecc"} err="failed to get container status \"884a7625c02fed00db769ed10f5c30e096f55ef9b572b71f5d3f38a42c7f5ecc\": rpc error: code = NotFound desc = could not find container \"884a7625c02fed00db769ed10f5c30e096f55ef9b572b71f5d3f38a42c7f5ecc\": container with ID starting with 884a7625c02fed00db769ed10f5c30e096f55ef9b572b71f5d3f38a42c7f5ecc not found: ID does not exist" Feb 02 15:50:12 crc kubenswrapper[4733]: I0202 15:50:12.045925 4733 scope.go:117] "RemoveContainer" containerID="fac49fae91c304d5ba93f5ac1352af229c9ae06429860065c7979d5bec4203c6" Feb 02 15:50:12 crc kubenswrapper[4733]: E0202 15:50:12.046702 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fac49fae91c304d5ba93f5ac1352af229c9ae06429860065c7979d5bec4203c6\": container with ID starting with fac49fae91c304d5ba93f5ac1352af229c9ae06429860065c7979d5bec4203c6 not found: ID does not exist" containerID="fac49fae91c304d5ba93f5ac1352af229c9ae06429860065c7979d5bec4203c6" Feb 02 15:50:12 crc kubenswrapper[4733]: I0202 15:50:12.046850 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fac49fae91c304d5ba93f5ac1352af229c9ae06429860065c7979d5bec4203c6"} err="failed to get container status \"fac49fae91c304d5ba93f5ac1352af229c9ae06429860065c7979d5bec4203c6\": rpc error: code = NotFound desc = could not find container \"fac49fae91c304d5ba93f5ac1352af229c9ae06429860065c7979d5bec4203c6\": container with ID starting with fac49fae91c304d5ba93f5ac1352af229c9ae06429860065c7979d5bec4203c6 not found: ID does not exist" Feb 02 15:50:12 crc kubenswrapper[4733]: I0202 15:50:12.046953 4733 scope.go:117] "RemoveContainer" containerID="3ac56d8ed4e4b3a07a44435c07e55954da4a7c4eb30f4cf55fd053881dd69a64" Feb 02 15:50:12 crc kubenswrapper[4733]: E0202 15:50:12.049663 4733 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ac56d8ed4e4b3a07a44435c07e55954da4a7c4eb30f4cf55fd053881dd69a64\": container with ID starting with 3ac56d8ed4e4b3a07a44435c07e55954da4a7c4eb30f4cf55fd053881dd69a64 not found: ID does not exist" containerID="3ac56d8ed4e4b3a07a44435c07e55954da4a7c4eb30f4cf55fd053881dd69a64" Feb 02 15:50:12 crc kubenswrapper[4733]: I0202 15:50:12.049711 4733 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ac56d8ed4e4b3a07a44435c07e55954da4a7c4eb30f4cf55fd053881dd69a64"} err="failed to get container status \"3ac56d8ed4e4b3a07a44435c07e55954da4a7c4eb30f4cf55fd053881dd69a64\": rpc error: code = NotFound desc = could not find container \"3ac56d8ed4e4b3a07a44435c07e55954da4a7c4eb30f4cf55fd053881dd69a64\": container with ID starting with 3ac56d8ed4e4b3a07a44435c07e55954da4a7c4eb30f4cf55fd053881dd69a64 not found: ID does not exist" Feb 02 15:50:13 crc kubenswrapper[4733]: I0202 15:50:13.270955 4733 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6498c564-0dc9-4c8b-aadf-a8d921c560c5" path="/var/lib/kubelet/pods/6498c564-0dc9-4c8b-aadf-a8d921c560c5/volumes" Feb 02 15:51:04 crc kubenswrapper[4733]: I0202 15:51:04.988225 4733 patch_prober.go:28] interesting pod/machine-config-daemon-zgsv9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 02 15:51:04 crc kubenswrapper[4733]: I0202 15:51:04.988734 4733 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zgsv9" podUID="83d0e810-ea5c-47aa-947b-7865b045bf94" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515140143770024447 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015140143771017365 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015140137302016501 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015140137302015451 5ustar corecore